var/home/core/zuul-output/0000755000175000017500000000000015067422761014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067446456015512 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005262340515067446447017725 0ustar rootrootOct 02 07:16:17 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 07:16:17 crc restorecon[4733]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:17 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:18 crc restorecon[4733]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 07:16:19 crc kubenswrapper[4960]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:19 crc kubenswrapper[4960]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 07:16:19 crc kubenswrapper[4960]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:19 crc kubenswrapper[4960]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:19 crc kubenswrapper[4960]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 07:16:19 crc kubenswrapper[4960]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.876466 4960 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.881921 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882031 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882047 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882063 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882075 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882089 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882104 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882116 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882128 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882140 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882163 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882174 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882219 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882230 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882240 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882252 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882264 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882276 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882286 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882297 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882307 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882362 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882579 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882606 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882617 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882626 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882635 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882644 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882652 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882666 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882680 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882692 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882759 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882770 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.882779 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883184 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883197 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883205 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883221 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883230 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883239 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883247 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883255 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883262 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883271 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883278 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883286 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883294 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883302 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883310 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883318 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883326 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883334 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883341 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883351 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883360 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883367 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883375 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883384 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883398 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883410 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883419 4960 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883429 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883443 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883453 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883462 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883470 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883478 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883485 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883492 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.883500 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883717 4960 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883737 4960 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883756 4960 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883768 4960 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883793 4960 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883812 4960 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883833 4960 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883846 4960 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883857 4960 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883867 4960 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883878 4960 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883889 4960 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883899 4960 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883908 4960 flags.go:64] FLAG: --cgroup-root="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883918 4960 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883931 4960 flags.go:64] FLAG: --client-ca-file="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883940 4960 flags.go:64] FLAG: --cloud-config="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883949 4960 flags.go:64] FLAG: --cloud-provider="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883958 4960 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.883971 4960 flags.go:64] FLAG: --cluster-domain="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884017 4960 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884027 4960 flags.go:64] FLAG: --config-dir="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884036 4960 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884046 4960 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884058 4960 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884069 4960 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884079 4960 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884088 4960 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884098 4960 flags.go:64] FLAG: --contention-profiling="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884107 4960 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884116 4960 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884126 4960 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884135 4960 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884148 4960 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884157 4960 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884166 4960 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884175 4960 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884185 4960 flags.go:64] FLAG: --enable-server="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884194 4960 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884207 4960 flags.go:64] FLAG: --event-burst="100" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884217 4960 flags.go:64] FLAG: --event-qps="50" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884226 4960 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884236 4960 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884245 4960 flags.go:64] FLAG: --eviction-hard="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884257 4960 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884266 4960 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884276 4960 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884286 4960 flags.go:64] FLAG: --eviction-soft="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884295 4960 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884304 4960 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884313 4960 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884322 4960 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884331 4960 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884340 4960 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884349 4960 flags.go:64] FLAG: --feature-gates="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884360 4960 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884369 4960 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884379 4960 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884388 4960 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884398 4960 flags.go:64] FLAG: --healthz-port="10248" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884408 4960 flags.go:64] FLAG: --help="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884418 4960 flags.go:64] FLAG: --hostname-override="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884426 4960 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884436 4960 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884445 4960 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884454 4960 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884463 4960 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884472 4960 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884481 4960 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884490 4960 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884499 4960 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884508 4960 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884517 4960 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884526 4960 flags.go:64] FLAG: --kube-reserved="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884536 4960 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884544 4960 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884553 4960 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884562 4960 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884575 4960 flags.go:64] FLAG: --lock-file="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884584 4960 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884593 4960 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884602 4960 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884618 4960 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884630 4960 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884639 4960 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884648 4960 flags.go:64] FLAG: --logging-format="text" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884659 4960 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884669 4960 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884678 4960 flags.go:64] FLAG: --manifest-url="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884687 4960 flags.go:64] FLAG: --manifest-url-header="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884712 4960 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884721 4960 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884733 4960 flags.go:64] FLAG: --max-pods="110" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884743 4960 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884754 4960 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884764 4960 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884774 4960 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884784 4960 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884794 4960 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884804 4960 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884833 4960 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884843 4960 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884852 4960 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884861 4960 flags.go:64] FLAG: --pod-cidr="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884870 4960 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884891 4960 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884926 4960 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884936 4960 flags.go:64] FLAG: --pods-per-core="0" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884945 4960 flags.go:64] FLAG: --port="10250" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884955 4960 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884964 4960 flags.go:64] FLAG: --provider-id="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.884973 4960 flags.go:64] FLAG: --qos-reserved="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885005 4960 flags.go:64] FLAG: --read-only-port="10255" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885015 4960 flags.go:64] FLAG: --register-node="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885024 4960 flags.go:64] FLAG: --register-schedulable="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885034 4960 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885050 4960 flags.go:64] FLAG: --registry-burst="10" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885059 4960 flags.go:64] FLAG: --registry-qps="5" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885068 4960 flags.go:64] FLAG: --reserved-cpus="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885079 4960 flags.go:64] FLAG: --reserved-memory="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885091 4960 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885101 4960 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885110 4960 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885119 4960 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885128 4960 flags.go:64] FLAG: --runonce="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885138 4960 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885148 4960 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885157 4960 flags.go:64] FLAG: --seccomp-default="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885168 4960 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885177 4960 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885187 4960 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885196 4960 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885206 4960 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885215 4960 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885224 4960 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885233 4960 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885242 4960 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885252 4960 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885262 4960 flags.go:64] FLAG: --system-cgroups="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885272 4960 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885289 4960 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885299 4960 flags.go:64] FLAG: --tls-cert-file="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885308 4960 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885320 4960 flags.go:64] FLAG: --tls-min-version="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885329 4960 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885338 4960 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885347 4960 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885357 4960 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885366 4960 flags.go:64] FLAG: --v="2" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885378 4960 flags.go:64] FLAG: --version="false" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885391 4960 flags.go:64] FLAG: --vmodule="" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885402 4960 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.885412 4960 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889121 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889139 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889151 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889163 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889174 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889183 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889193 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889204 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889213 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889222 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889231 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889239 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889247 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889255 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889264 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889271 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889279 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889286 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889294 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889302 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889310 4960 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889317 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889325 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889333 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889341 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889348 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889356 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889364 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889372 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889379 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889387 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889395 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889402 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889410 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889418 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889425 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889433 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889440 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889449 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889457 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889468 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889479 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889487 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889495 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889503 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889512 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889520 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889528 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889536 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889543 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889551 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889559 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889567 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889575 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889585 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889593 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889601 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889608 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889616 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889624 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889632 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889640 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889647 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889654 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889662 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889669 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889680 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889688 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889696 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889703 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.889712 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.889740 4960 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.931408 4960 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.931463 4960 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931616 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931630 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931639 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931651 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931660 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931673 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931685 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931695 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931705 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931715 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931724 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931733 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931742 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931751 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931760 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931768 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931779 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931789 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931798 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931809 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931819 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931829 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931839 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931848 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931860 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931870 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931880 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931889 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931898 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931909 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931917 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931926 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931934 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931943 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931952 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931960 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.931968 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932004 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932014 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932024 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932032 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932041 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932049 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932057 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932066 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932074 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932082 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932094 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932106 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932117 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932128 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932138 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932150 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932161 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932173 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932184 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932195 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932206 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932216 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932227 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932238 4960 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932249 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932259 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932269 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932277 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932289 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932298 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932307 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932315 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932325 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932333 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.932349 4960 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932591 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932605 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932616 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932625 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932634 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932644 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932652 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932661 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932672 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932684 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932695 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932707 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932719 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932730 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932742 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932754 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932764 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932775 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932786 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932800 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932817 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932831 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932843 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932854 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932865 4960 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932876 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932887 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932898 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932907 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932919 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932927 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932936 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932944 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932952 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932961 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.932980 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933019 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933028 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933038 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933048 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933060 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933069 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933079 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933089 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933098 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933106 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933115 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933123 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933132 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933140 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933148 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933157 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933166 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933175 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933183 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933191 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933200 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933208 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933217 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933225 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933233 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933241 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933250 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933259 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933267 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933296 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933305 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933314 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933322 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933331 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:19 crc kubenswrapper[4960]: W1002 07:16:19.933339 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.933354 4960 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.942387 4960 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.958868 4960 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.959073 4960 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.961198 4960 server.go:997] "Starting client certificate rotation" Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.961240 4960 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.962662 4960 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-10 01:29:33.154951601 +0000 UTC Oct 02 07:16:19 crc kubenswrapper[4960]: I1002 07:16:19.962849 4960 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1650h13m13.192110301s for next certificate rotation Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.044142 4960 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.046443 4960 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.079767 4960 log.go:25] "Validated CRI v1 runtime API" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.159596 4960 log.go:25] "Validated CRI v1 image API" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.162035 4960 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.176643 4960 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-07-11-38-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.176708 4960 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.212587 4960 manager.go:217] Machine: {Timestamp:2025-10-02 07:16:20.206265231 +0000 UTC m=+1.238211598 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:82b394c9-1025-4b62-b6ab-7caff0ff113f BootID:903088be-089c-4b2a-9679-59229550832d Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b8:4a:02 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b8:4a:02 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:17:90:64 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:74:f6:d7 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d8:8e:03 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f8:a5:ca Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:85:2e:fd Speed:-1 Mtu:1496} {Name:eth10 MacAddress:c6:db:22:88:22:5d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1e:bf:3d:7c:f3:8e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.213124 4960 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.213379 4960 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.215049 4960 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.215386 4960 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.215461 4960 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.215842 4960 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.215864 4960 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.216652 4960 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.216722 4960 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.217047 4960 state_mem.go:36] "Initialized new in-memory state store" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.217631 4960 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.228052 4960 kubelet.go:418] "Attempting to sync node with API server" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.228094 4960 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.228138 4960 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.228166 4960 kubelet.go:324] "Adding apiserver pod source" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.228186 4960 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.235898 4960 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.237678 4960 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 07:16:20 crc kubenswrapper[4960]: W1002 07:16:20.239770 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:20 crc kubenswrapper[4960]: W1002 07:16:20.239764 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.240137 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.240028 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.241369 4960 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244196 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244225 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244235 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244245 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244261 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244276 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244287 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244302 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244314 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244326 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244342 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.244354 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.246474 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.247185 4960 server.go:1280] "Started kubelet" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.249261 4960 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 07:16:20 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.249261 4960 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.258161 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.260070 4960 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.262195 4960 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.262261 4960 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.262332 4960 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.262355 4960 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.262762 4960 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.262852 4960 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.263116 4960 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 23:57:13.030364591 +0000 UTC Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.263262 4960 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1912h40m52.767111016s for next certificate rotation Oct 02 07:16:20 crc kubenswrapper[4960]: W1002 07:16:20.264649 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.264902 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.265473 4960 server.go:460] "Adding debug handlers to kubelet server" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.267453 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="200ms" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.261884 4960 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a9b51d739b44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 07:16:20.247139403 +0000 UTC m=+1.279085700,LastTimestamp:2025-10-02 07:16:20.247139403 +0000 UTC m=+1.279085700,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.273676 4960 factory.go:153] Registering CRI-O factory Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.274214 4960 factory.go:221] Registration of the crio container factory successfully Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.274613 4960 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.274806 4960 factory.go:55] Registering systemd factory Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.275015 4960 factory.go:221] Registration of the systemd container factory successfully Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.275212 4960 factory.go:103] Registering Raw factory Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.275487 4960 manager.go:1196] Started watching for new ooms in manager Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.277606 4960 manager.go:319] Starting recovery of all containers Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285616 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285719 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285736 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285750 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285763 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285775 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285787 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285799 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285813 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285826 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285839 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285856 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285868 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285884 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285898 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285911 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285924 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285936 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285947 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285963 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.285978 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286071 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286087 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286100 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286119 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286165 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286185 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286203 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286219 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286237 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286253 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286274 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286313 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286330 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286347 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286366 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286383 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286397 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286410 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286424 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286441 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286458 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286473 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286491 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286505 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286519 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286534 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286556 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286577 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286594 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286610 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286626 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286651 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286672 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286692 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286712 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286731 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286747 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286765 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286782 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286799 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286816 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286833 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286852 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286870 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286886 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286902 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286922 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286937 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.286957 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287003 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287021 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287040 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287054 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287067 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287080 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287092 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287104 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287117 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287133 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287145 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287159 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287172 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287184 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287196 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287209 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287225 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287238 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287250 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287263 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287277 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287290 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287301 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287313 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287330 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287342 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287354 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287368 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287379 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287392 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287405 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287417 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287429 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287440 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287464 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287478 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287492 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287507 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287521 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287534 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287547 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287563 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287576 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287589 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287602 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287614 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287626 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287639 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287652 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287663 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287678 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287689 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287701 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287713 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287725 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287737 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287752 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287766 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287778 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287790 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287801 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287814 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287828 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287841 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287853 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287866 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287880 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287893 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287905 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287917 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287931 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287943 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287955 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.287970 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288004 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288016 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288028 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288040 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288052 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288066 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288078 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288090 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.288103 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290710 4960 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290789 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290817 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290839 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290864 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290884 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290910 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290929 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.290951 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291007 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291029 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291049 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291068 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291088 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291110 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291129 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291147 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291169 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291189 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291208 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291232 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291251 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291271 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291310 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291330 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291349 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291368 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291386 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291403 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291422 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291446 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291466 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291523 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291545 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291565 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291589 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291609 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291631 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291652 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291671 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291690 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291759 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291781 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291802 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291822 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.291873 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292430 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292512 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292535 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292559 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292580 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292600 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292618 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292634 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292650 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292664 4960 reconstruct.go:97] "Volume reconstruction finished" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.292675 4960 reconciler.go:26] "Reconciler: start to sync state" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.307124 4960 manager.go:324] Recovery completed Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.324108 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.325613 4960 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.326601 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.326643 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.326657 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.327719 4960 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.327743 4960 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.327771 4960 state_mem.go:36] "Initialized new in-memory state store" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.328519 4960 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.328569 4960 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.328602 4960 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.328663 4960 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 07:16:20 crc kubenswrapper[4960]: W1002 07:16:20.329733 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.329887 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.363640 4960 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.396878 4960 policy_none.go:49] "None policy: Start" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.398496 4960 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.398554 4960 state_mem.go:35] "Initializing new in-memory state store" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.428965 4960 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.464516 4960 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.466815 4960 manager.go:334] "Starting Device Plugin manager" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.466885 4960 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.466901 4960 server.go:79] "Starting device plugin registration server" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.467571 4960 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.467593 4960 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.467809 4960 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.468070 4960 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.468090 4960 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.468477 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="400ms" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.479907 4960 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.568655 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.570932 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.571036 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.571055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.571128 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.572300 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.629257 4960 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.629500 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.632136 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.632180 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.632229 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.632466 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.632917 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.633008 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.634742 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.634789 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.634799 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.634726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.634910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.634920 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.635029 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.635223 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.635271 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.636836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.636918 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.636936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.636934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.637032 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.637050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.637326 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.637452 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.637501 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.638745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.638782 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.638795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.638815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.638834 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.638867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.639021 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.639218 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.639294 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.639809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.639840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.639857 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.640161 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.640198 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.641114 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.641139 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.641136 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.641150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.641188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.641219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.697859 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.697926 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.697994 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698040 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698123 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698235 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698299 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698338 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698400 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698456 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.698943 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.699031 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.699075 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.699125 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.699162 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.773485 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.775331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.775383 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.775400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.775439 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.776066 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800644 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800738 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800788 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800822 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800857 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800890 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800922 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800926 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801012 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.800955 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801084 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801110 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801167 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801170 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801228 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801235 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801283 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801245 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801334 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801363 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801388 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801419 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801419 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801444 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801495 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801543 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801547 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801464 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801464 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.801709 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: E1002 07:16:20.870142 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="800ms" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.975936 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:20 crc kubenswrapper[4960]: I1002 07:16:20.985532 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.009292 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.023375 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.032215 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.086154 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-274785de93c8321eb9d52626d69cd0f7c15041cf97ea03a1ce9eb0dd981854d8 WatchSource:0}: Error finding container 274785de93c8321eb9d52626d69cd0f7c15041cf97ea03a1ce9eb0dd981854d8: Status 404 returned error can't find the container with id 274785de93c8321eb9d52626d69cd0f7c15041cf97ea03a1ce9eb0dd981854d8 Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.092882 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-40cf50c569a0269fb8c5f73919f568bfdd9df7fa39e60a001f9cc1711a66ecfb WatchSource:0}: Error finding container 40cf50c569a0269fb8c5f73919f568bfdd9df7fa39e60a001f9cc1711a66ecfb: Status 404 returned error can't find the container with id 40cf50c569a0269fb8c5f73919f568bfdd9df7fa39e60a001f9cc1711a66ecfb Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.096746 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-e2ec28f958956af003d69354562626e0cf289cf4ddf1f02fa8efc15efa23bf0a WatchSource:0}: Error finding container e2ec28f958956af003d69354562626e0cf289cf4ddf1f02fa8efc15efa23bf0a: Status 404 returned error can't find the container with id e2ec28f958956af003d69354562626e0cf289cf4ddf1f02fa8efc15efa23bf0a Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.098311 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e58e83552288818ee662821da301ac166037e6bc1f6e9f3abdc54ce0f31c823e WatchSource:0}: Error finding container e58e83552288818ee662821da301ac166037e6bc1f6e9f3abdc54ce0f31c823e: Status 404 returned error can't find the container with id e58e83552288818ee662821da301ac166037e6bc1f6e9f3abdc54ce0f31c823e Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.104769 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-710beee7f6301e99ab5f748ce64b7bd347b4c3b1f14d16d9fae27ea557addcb4 WatchSource:0}: Error finding container 710beee7f6301e99ab5f748ce64b7bd347b4c3b1f14d16d9fae27ea557addcb4: Status 404 returned error can't find the container with id 710beee7f6301e99ab5f748ce64b7bd347b4c3b1f14d16d9fae27ea557addcb4 Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.177184 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.179423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.179496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.179516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.179568 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:21 crc kubenswrapper[4960]: E1002 07:16:21.180458 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.184576 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:21 crc kubenswrapper[4960]: E1002 07:16:21.184722 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.199096 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:21 crc kubenswrapper[4960]: E1002 07:16:21.199176 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.199374 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:21 crc kubenswrapper[4960]: E1002 07:16:21.199534 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.259845 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.334372 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"274785de93c8321eb9d52626d69cd0f7c15041cf97ea03a1ce9eb0dd981854d8"} Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.337428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e58e83552288818ee662821da301ac166037e6bc1f6e9f3abdc54ce0f31c823e"} Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.340258 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"710beee7f6301e99ab5f748ce64b7bd347b4c3b1f14d16d9fae27ea557addcb4"} Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.342126 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e2ec28f958956af003d69354562626e0cf289cf4ddf1f02fa8efc15efa23bf0a"} Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.343518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"40cf50c569a0269fb8c5f73919f568bfdd9df7fa39e60a001f9cc1711a66ecfb"} Oct 02 07:16:21 crc kubenswrapper[4960]: W1002 07:16:21.392170 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:21 crc kubenswrapper[4960]: E1002 07:16:21.392334 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:21 crc kubenswrapper[4960]: E1002 07:16:21.671888 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="1.6s" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.981587 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.985475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.985553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.985579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:21 crc kubenswrapper[4960]: I1002 07:16:21.985626 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:21 crc kubenswrapper[4960]: E1002 07:16:21.986499 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.259746 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.349644 4960 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="05963d1e29ac23bbd6483b6bc1876a3ec39feb299e4d9aaa6813aab0a959f7b0" exitCode=0 Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.349791 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"05963d1e29ac23bbd6483b6bc1876a3ec39feb299e4d9aaa6813aab0a959f7b0"} Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.350033 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.351467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.351509 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.351534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.356146 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff"} Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.356201 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960"} Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.359392 4960 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33" exitCode=0 Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.359452 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33"} Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.359531 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.360734 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.360812 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.360833 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.361816 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff" exitCode=0 Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.361867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff"} Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.361955 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.363409 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.363448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.363461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.364189 4960 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529" exitCode=0 Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.364232 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529"} Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.364362 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.365270 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.365827 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.365867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.365882 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.366084 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.366113 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:22 crc kubenswrapper[4960]: I1002 07:16:22.366126 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.259293 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:23 crc kubenswrapper[4960]: E1002 07:16:23.272650 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="3.2s" Oct 02 07:16:23 crc kubenswrapper[4960]: W1002 07:16:23.373940 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:23 crc kubenswrapper[4960]: E1002 07:16:23.374079 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.374568 4960 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe" exitCode=0 Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.374684 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.374707 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.375931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.376018 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.376033 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.378166 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"89e0a99040768369fd7eb9c9c96a44ee6eae13519e5d153135fa5e05d4882023"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.378198 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.379692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.379736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.379750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.382797 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.382852 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.382868 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.384236 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.384261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.384273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.388696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.388740 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.388753 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.388802 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.390103 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.390176 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.390190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.403203 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.403243 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.403261 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd"} Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.403276 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f"} Oct 02 07:16:23 crc kubenswrapper[4960]: W1002 07:16:23.456252 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:23 crc kubenswrapper[4960]: E1002 07:16:23.456387 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:23 crc kubenswrapper[4960]: W1002 07:16:23.517484 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Oct 02 07:16:23 crc kubenswrapper[4960]: E1002 07:16:23.517571 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.587640 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.589051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.589104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.589121 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:23 crc kubenswrapper[4960]: I1002 07:16:23.589159 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:23 crc kubenswrapper[4960]: E1002 07:16:23.589810 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.410527 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e"} Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.410725 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.412605 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.412650 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.412667 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.413789 4960 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407" exitCode=0 Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.413855 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407"} Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.413907 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.413963 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.414028 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.414101 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.413914 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416116 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416083 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416854 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.416965 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:24 crc kubenswrapper[4960]: I1002 07:16:24.703943 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.422737 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908"} Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.422817 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d"} Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.422837 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c"} Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.422853 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.423060 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.423183 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.428288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.428291 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.428321 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.428483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.428408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:25 crc kubenswrapper[4960]: I1002 07:16:25.428575 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.431509 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d"} Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.431622 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806"} Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.431578 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.431743 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.433353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.433411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.433435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.433676 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.433719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.433731 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.632300 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.632569 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.634428 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.634522 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.634536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.790087 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.791570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.791604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.791618 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:26 crc kubenswrapper[4960]: I1002 07:16:26.791647 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.150915 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.151236 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.152795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.152855 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.152870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.436165 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.437342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.437395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.437407 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:27 crc kubenswrapper[4960]: I1002 07:16:27.830586 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 07:16:28 crc kubenswrapper[4960]: I1002 07:16:28.441778 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:28 crc kubenswrapper[4960]: I1002 07:16:28.442861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:28 crc kubenswrapper[4960]: I1002 07:16:28.442909 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:28 crc kubenswrapper[4960]: I1002 07:16:28.442921 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.153255 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.153704 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.155628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.155881 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.156149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.317905 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.450544 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.452762 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.452809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:29 crc kubenswrapper[4960]: I1002 07:16:29.452822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:30 crc kubenswrapper[4960]: I1002 07:16:30.151040 4960 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 07:16:30 crc kubenswrapper[4960]: I1002 07:16:30.151190 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:16:30 crc kubenswrapper[4960]: E1002 07:16:30.480247 4960 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 07:16:31 crc kubenswrapper[4960]: I1002 07:16:31.760639 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 07:16:31 crc kubenswrapper[4960]: I1002 07:16:31.760898 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:31 crc kubenswrapper[4960]: I1002 07:16:31.762427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:31 crc kubenswrapper[4960]: I1002 07:16:31.762459 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:31 crc kubenswrapper[4960]: I1002 07:16:31.762467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.216476 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.216946 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.218870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.218936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.218960 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.225498 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.460170 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.460415 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.461727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.461807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.461826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:32 crc kubenswrapper[4960]: I1002 07:16:32.471636 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:33 crc kubenswrapper[4960]: I1002 07:16:33.463667 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:33 crc kubenswrapper[4960]: I1002 07:16:33.465395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:33 crc kubenswrapper[4960]: I1002 07:16:33.465451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:33 crc kubenswrapper[4960]: I1002 07:16:33.465469 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:33 crc kubenswrapper[4960]: I1002 07:16:33.851826 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 02 07:16:33 crc kubenswrapper[4960]: I1002 07:16:33.851904 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.009012 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.009105 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 02 07:16:34 crc kubenswrapper[4960]: W1002 07:16:34.240617 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.240760 4960 trace.go:236] Trace[1283505381]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:24.238) (total time: 10002ms): Oct 02 07:16:34 crc kubenswrapper[4960]: Trace[1283505381]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:16:34.240) Oct 02 07:16:34 crc kubenswrapper[4960]: Trace[1283505381]: [10.002100959s] [10.002100959s] END Oct 02 07:16:34 crc kubenswrapper[4960]: E1002 07:16:34.240791 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.259638 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.467575 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.468722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.468750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.468762 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.575128 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.575262 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.579879 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 07:16:34 crc kubenswrapper[4960]: I1002 07:16:34.580037 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 07:16:37 crc kubenswrapper[4960]: I1002 07:16:37.693794 4960 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.236815 4960 apiserver.go:52] "Watching apiserver" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.244479 4960 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.244882 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.245439 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.245485 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:38 crc kubenswrapper[4960]: E1002 07:16:38.245587 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.245454 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.246023 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.246052 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.246092 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:38 crc kubenswrapper[4960]: E1002 07:16:38.246658 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:38 crc kubenswrapper[4960]: E1002 07:16:38.246909 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.249188 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.249221 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.249244 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.249226 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.250023 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.250472 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.250822 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.251097 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.252043 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.264871 4960 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.328412 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.349768 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.368233 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.385840 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.401803 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.415208 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.431366 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:38 crc kubenswrapper[4960]: I1002 07:16:38.451335 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.159700 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.168355 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.182873 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.183547 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.200189 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.219369 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.234940 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.246992 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.263925 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.281224 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.298439 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.309773 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.327034 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.329153 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.329327 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.340780 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.354310 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.366533 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.490250 4960 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.575075 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.578043 4960 trace.go:236] Trace[993111043]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:27.083) (total time: 12494ms): Oct 02 07:16:39 crc kubenswrapper[4960]: Trace[993111043]: ---"Objects listed" error: 12494ms (07:16:39.577) Oct 02 07:16:39 crc kubenswrapper[4960]: Trace[993111043]: [12.494694087s] [12.494694087s] END Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.578066 4960 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.578187 4960 trace.go:236] Trace[90791471]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:27.210) (total time: 12367ms): Oct 02 07:16:39 crc kubenswrapper[4960]: Trace[90791471]: ---"Objects listed" error: 12367ms (07:16:39.578) Oct 02 07:16:39 crc kubenswrapper[4960]: Trace[90791471]: [12.367212903s] [12.367212903s] END Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.578223 4960 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.578232 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.578548 4960 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.580004 4960 trace.go:236] Trace[717086254]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:27.422) (total time: 12157ms): Oct 02 07:16:39 crc kubenswrapper[4960]: Trace[717086254]: ---"Objects listed" error: 12157ms (07:16:39.579) Oct 02 07:16:39 crc kubenswrapper[4960]: Trace[717086254]: [12.157584379s] [12.157584379s] END Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.580031 4960 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.616272 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.631598 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.646198 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.674114 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.678800 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.678844 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.678871 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.678899 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.678931 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.678958 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679007 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679034 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679068 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679094 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679121 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679147 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679171 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679198 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679219 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679243 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679275 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679316 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679341 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679372 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679403 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679431 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679466 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679495 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679518 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679542 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679565 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679595 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679618 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679642 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679666 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679689 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679713 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679736 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679759 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679785 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679808 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679830 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679851 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679872 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679894 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679914 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679935 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.679970 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680012 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680036 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680060 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680081 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680105 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680128 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680151 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680174 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680197 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680219 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680242 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680264 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680286 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680307 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680330 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680351 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680372 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680394 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680416 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680445 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680467 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680494 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680517 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680539 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680564 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681094 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681103 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681148 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681305 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681447 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681460 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681517 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681537 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681718 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681832 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681928 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681950 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681931 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.681940 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682034 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682088 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682168 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682196 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682228 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682357 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682427 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682414 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682474 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682450 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682499 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682592 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682636 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682712 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682733 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682795 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682807 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682843 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682904 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682940 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.682959 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683002 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683073 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683129 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683188 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683284 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683334 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683355 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683540 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683568 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683591 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683616 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683699 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683737 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.683949 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684007 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684015 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.680794 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684114 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684146 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684171 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684197 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684225 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684247 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684273 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684297 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684324 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684347 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684368 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684390 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684410 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684419 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684433 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684450 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684457 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684520 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684557 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684581 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.684605 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691077 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691196 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691232 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691262 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691288 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691311 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691337 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691364 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691388 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691418 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691444 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691478 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691502 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691531 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691555 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691577 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691602 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691627 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691650 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691672 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691694 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691717 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691742 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691769 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691792 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691814 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691840 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691862 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691886 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691906 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691932 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691965 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692006 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692031 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692057 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692082 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692100 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692126 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692150 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692168 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692191 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692217 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692237 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692260 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692284 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692305 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692326 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692349 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692384 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692408 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692429 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692453 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692474 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692501 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692525 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692550 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692571 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692598 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695352 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695419 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695451 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695473 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695501 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695526 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695547 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695569 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695593 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695616 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.688324 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.689088 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.689294 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.689351 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.689643 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.689852 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.690473 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.690488 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691136 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691203 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691372 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.691421 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.692633 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.693076 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.694738 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.694768 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695625 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.696106 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.696440 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.696583 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.696919 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.697228 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.697731 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.699444 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.699703 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.699767 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.700263 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.700540 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.700712 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.700882 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.701158 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.701323 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.701704 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.701724 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.701895 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.702480 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.702768 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.702777 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.702836 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.705700 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.705940 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.706184 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.706204 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.706278 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.706553 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.706579 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.707344 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.707674 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.707728 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.707760 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.707962 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.708061 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.708186 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.708200 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.708374 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.695638 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.708600 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.708716 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709223 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709396 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709688 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709738 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709768 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709803 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709824 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709851 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709877 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709895 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709917 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709941 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.709963 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710033 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710059 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710079 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710097 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710116 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710136 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710159 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710179 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710197 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710217 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710239 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710258 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710280 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710297 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710321 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710352 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710377 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710396 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710421 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710439 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710466 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710712 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.710861 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.711080 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.711198 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.711912 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.712830 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.712965 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715195 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715241 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715269 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715298 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715326 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715776 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.715902 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.716160 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.716235 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.716269 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.714902 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.716104 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.716841 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.716911 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717045 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717108 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717136 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.716655 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717196 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717222 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717252 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717293 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717379 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717396 4960 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717392 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.717459 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.717433 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:16:40.217400553 +0000 UTC m=+21.249346840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.718365 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.719422 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717601 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.717932 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.718009 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.718107 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.718347 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.719859 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.718479 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.719888 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.719914 4960 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.718650 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.719350 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.720445 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:40.220421618 +0000 UTC m=+21.252367905 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.719704 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.720702 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.721281 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.721661 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.721715 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.721777 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.721792 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.721805 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.721815 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722050 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722110 4960 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722150 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722176 4960 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722196 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722225 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722245 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722264 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722295 4960 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.722341 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:40.222305055 +0000 UTC m=+21.254251542 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722370 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722474 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722494 4960 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722512 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722537 4960 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722550 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722565 4960 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722579 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722593 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722606 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722622 4960 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722637 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722631 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722647 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722677 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722692 4960 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722705 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722717 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722729 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722741 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722752 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722765 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722777 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722787 4960 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.722799 4960 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723451 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723463 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723476 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723487 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723499 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723514 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723527 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723539 4960 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723551 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723562 4960 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723573 4960 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723585 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723595 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723608 4960 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723632 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723643 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723653 4960 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723665 4960 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723675 4960 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723685 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723696 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723707 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723717 4960 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723727 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723740 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723752 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723815 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723828 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723841 4960 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723853 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723865 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723875 4960 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723888 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723900 4960 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723911 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723921 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723933 4960 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723947 4960 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.723957 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724128 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724142 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724153 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724164 4960 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724175 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724185 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724196 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724206 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724217 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724227 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724241 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724252 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724264 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724274 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724284 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724294 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724304 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724315 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724324 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724335 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724345 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724355 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724370 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724387 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724398 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724411 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724423 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724433 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724443 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724453 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724463 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724472 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724483 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724494 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724503 4960 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724513 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724522 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724532 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724543 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724553 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724563 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724573 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724584 4960 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724595 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724606 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724616 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.724929 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.725075 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.725101 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.725142 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.725234 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.725595 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.725628 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.725672 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.728537 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.729895 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.729999 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.730107 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.730143 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.730161 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.730257 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:40.230228423 +0000 UTC m=+21.262174720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.731094 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.731309 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.731422 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.731785 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.731834 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.732131 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.732281 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.732368 4960 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.732575 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.732884 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.733449 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.733504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.733704 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.733786 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.734012 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.734043 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.734063 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.734059 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: E1002 07:16:39.734133 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:40.23410765 +0000 UTC m=+21.266053937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.734664 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.734778 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.735039 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.735528 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.735810 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.735892 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.736033 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.736219 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.736511 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.736966 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737119 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737208 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737319 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737489 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737591 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737633 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737698 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.737858 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.738071 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.738338 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.739673 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.739722 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.740167 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.740334 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.740896 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.741660 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.743013 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.743111 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.743563 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.747092 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.748665 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.754666 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.766463 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.773187 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.775839 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.781005 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.804056 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825065 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825110 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825147 4960 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825159 4960 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825187 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825198 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825206 4960 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825215 4960 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825224 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825232 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825242 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825251 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825259 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825267 4960 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825276 4960 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825302 4960 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825313 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825323 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825332 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825341 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825350 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825359 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825369 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825377 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825386 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825402 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825411 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825419 4960 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825428 4960 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825437 4960 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825448 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825458 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825469 4960 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825478 4960 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825488 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825497 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825505 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825513 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825521 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825532 4960 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825541 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825550 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825559 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825568 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825576 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825585 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825593 4960 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825602 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825610 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825619 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825630 4960 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825638 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825646 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825654 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825662 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825685 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825696 4960 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825710 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825720 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825729 4960 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825738 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825750 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825759 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825768 4960 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825776 4960 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825784 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825795 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825803 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825811 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825819 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825827 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825836 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825844 4960 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.825891 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:39 crc kubenswrapper[4960]: I1002 07:16:39.826087 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.065426 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:16:40 crc kubenswrapper[4960]: W1002 07:16:40.076270 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-da075ce0baf3c502c6ee8625526e4d9220db833539cacc7cf0df9f7bc767ee76 WatchSource:0}: Error finding container da075ce0baf3c502c6ee8625526e4d9220db833539cacc7cf0df9f7bc767ee76: Status 404 returned error can't find the container with id da075ce0baf3c502c6ee8625526e4d9220db833539cacc7cf0df9f7bc767ee76 Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.113270 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:16:40 crc kubenswrapper[4960]: W1002 07:16:40.127845 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-08830300aba32952298a918d6f29fd66515392b63d045e07f3f99500d4e77aeb WatchSource:0}: Error finding container 08830300aba32952298a918d6f29fd66515392b63d045e07f3f99500d4e77aeb: Status 404 returned error can't find the container with id 08830300aba32952298a918d6f29fd66515392b63d045e07f3f99500d4e77aeb Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.151270 4960 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.151401 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.229095 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.229190 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.229216 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.229367 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.229425 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:41.22940915 +0000 UTC m=+22.261355437 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.229487 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:16:41.229481262 +0000 UTC m=+22.261427549 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.229519 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.229540 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:41.229534074 +0000 UTC m=+22.261480361 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.329771 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.330062 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330163 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.330246 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330280 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.330327 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330530 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330587 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330530 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330615 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330644 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330667 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330725 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:41.33069349 +0000 UTC m=+22.362639817 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:40 crc kubenswrapper[4960]: E1002 07:16:40.330835 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:41.330815673 +0000 UTC m=+22.362762000 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.334523 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.335252 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.337001 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.337787 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.339216 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.339961 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.340758 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.340495 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.342000 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.342723 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.344105 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.344727 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.346780 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.347486 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.348541 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.349274 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.350013 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.351949 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.352565 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.353595 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.355048 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.355668 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.356969 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.357556 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.359107 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.359890 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.360599 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.361592 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.362138 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.362753 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.363347 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.363835 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.363896 4960 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.364022 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.365508 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.366069 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.366570 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.372307 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.374437 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.375022 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.376189 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.376863 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.377342 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.378447 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.379559 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.380075 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.380219 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.381078 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.381627 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.382590 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.383586 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.384498 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.384957 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.385482 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.386600 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.387191 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.389165 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.395407 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.418470 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.430951 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.441939 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.483634 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"08830300aba32952298a918d6f29fd66515392b63d045e07f3f99500d4e77aeb"} Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.484805 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3"} Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.484871 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"da075ce0baf3c502c6ee8625526e4d9220db833539cacc7cf0df9f7bc767ee76"} Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.486298 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980"} Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.486336 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3"} Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.486351 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4071d2f56093d8830a5ceef8233a7934b611dba62224d146f096c52251c493b7"} Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.506596 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.523586 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.538501 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.551757 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.569931 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.584708 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.609372 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.632500 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.648859 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.663943 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.676871 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.690355 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.704688 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:40 crc kubenswrapper[4960]: I1002 07:16:40.719477 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.238945 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.239642 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:16:43.239582459 +0000 UTC m=+24.271528746 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.240424 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.240492 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.240701 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.240790 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:43.240755299 +0000 UTC m=+24.272701586 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.240865 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.240912 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:43.240901502 +0000 UTC m=+24.272847799 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.329324 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.329506 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.341676 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.341731 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.341879 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.341902 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.341916 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.341921 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.341952 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.341969 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.341990 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:43.341956076 +0000 UTC m=+24.373902363 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:41 crc kubenswrapper[4960]: E1002 07:16:41.342067 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:43.342032138 +0000 UTC m=+24.373978435 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.519437 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-bqvdt"] Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.519757 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.524022 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.528966 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.539943 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.553729 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.607584 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.642583 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.645278 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xq84\" (UniqueName: \"kubernetes.io/projected/e523ebca-7bf0-4904-9671-8c0381b01be8-kube-api-access-6xq84\") pod \"node-resolver-bqvdt\" (UID: \"e523ebca-7bf0-4904-9671-8c0381b01be8\") " pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.645353 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e523ebca-7bf0-4904-9671-8c0381b01be8-hosts-file\") pod \"node-resolver-bqvdt\" (UID: \"e523ebca-7bf0-4904-9671-8c0381b01be8\") " pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.682560 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.721061 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.745742 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xq84\" (UniqueName: \"kubernetes.io/projected/e523ebca-7bf0-4904-9671-8c0381b01be8-kube-api-access-6xq84\") pod \"node-resolver-bqvdt\" (UID: \"e523ebca-7bf0-4904-9671-8c0381b01be8\") " pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.745809 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e523ebca-7bf0-4904-9671-8c0381b01be8-hosts-file\") pod \"node-resolver-bqvdt\" (UID: \"e523ebca-7bf0-4904-9671-8c0381b01be8\") " pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.745890 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e523ebca-7bf0-4904-9671-8c0381b01be8-hosts-file\") pod \"node-resolver-bqvdt\" (UID: \"e523ebca-7bf0-4904-9671-8c0381b01be8\") " pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.754793 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.767069 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.775603 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xq84\" (UniqueName: \"kubernetes.io/projected/e523ebca-7bf0-4904-9671-8c0381b01be8-kube-api-access-6xq84\") pod \"node-resolver-bqvdt\" (UID: \"e523ebca-7bf0-4904-9671-8c0381b01be8\") " pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.788706 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.803741 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.823744 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.824230 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.832630 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bqvdt" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.840057 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: W1002 07:16:41.853109 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode523ebca_7bf0_4904_9671_8c0381b01be8.slice/crio-e23ef63206138c8d4a71f442fdd966336dbdc2b3f4113bbeeae0ddedf51fdebd WatchSource:0}: Error finding container e23ef63206138c8d4a71f442fdd966336dbdc2b3f4113bbeeae0ddedf51fdebd: Status 404 returned error can't find the container with id e23ef63206138c8d4a71f442fdd966336dbdc2b3f4113bbeeae0ddedf51fdebd Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.865875 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.895327 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.912842 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.920218 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.929731 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.945112 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.963084 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.977302 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:41 crc kubenswrapper[4960]: I1002 07:16:41.993568 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:41Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.010957 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.043786 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.058124 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.083392 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.100532 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.116113 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.142308 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.329395 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.329460 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:42 crc kubenswrapper[4960]: E1002 07:16:42.329575 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:42 crc kubenswrapper[4960]: E1002 07:16:42.329674 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.455437 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rfcbg"] Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.455951 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2dqxx"] Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.456256 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.456282 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.458228 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8dmqg"] Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.458931 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vm98j"] Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.461187 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.461663 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.461891 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.462003 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.462064 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.462191 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.462158 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.462402 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.463189 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.463200 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.463414 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.463568 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.468505 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.468523 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.468549 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.468698 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.468805 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.468838 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.469040 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.469883 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.470459 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.486847 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.493501 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bqvdt" event={"ID":"e523ebca-7bf0-4904-9671-8c0381b01be8","Type":"ContainerStarted","Data":"411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586"} Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.493630 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bqvdt" event={"ID":"e523ebca-7bf0-4904-9671-8c0381b01be8","Type":"ContainerStarted","Data":"e23ef63206138c8d4a71f442fdd966336dbdc2b3f4113bbeeae0ddedf51fdebd"} Oct 02 07:16:42 crc kubenswrapper[4960]: E1002 07:16:42.502536 4960 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.511307 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.524447 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.536256 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.547159 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.556800 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-cnibin\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.556844 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-systemd-units\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.556874 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-os-release\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.556895 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-cnibin\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.556914 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-socket-dir-parent\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557058 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-bin\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557094 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-etc-kubernetes\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557121 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxd2c\" (UniqueName: \"kubernetes.io/projected/41ab5426-0159-4610-b378-f4c2e20b82b9-kube-api-access-dxd2c\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557143 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-log-socket\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557176 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-daemon-config\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557274 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-var-lib-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-k8s-cni-cncf-io\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557414 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-systemd\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557454 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-etc-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557473 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-script-lib\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-system-cni-dir\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557569 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-system-cni-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557616 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-hostroot\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557730 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-cni-bin\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557853 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-cni-multus\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557901 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-slash\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.557931 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-ovn-kubernetes\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558056 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-multus-certs\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558095 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-node-log\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558086 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558124 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-os-release\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558151 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-kubelet\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558191 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-netns\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558216 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdwn2\" (UniqueName: \"kubernetes.io/projected/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-kube-api-access-kdwn2\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558242 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-netns\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558275 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a813e5e1-abad-433b-a1b3-24cab8119638-proxy-tls\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558311 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558341 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n84cc\" (UniqueName: \"kubernetes.io/projected/eb4216ca-0206-4794-bb05-efabd0dc96cb-kube-api-access-n84cc\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558371 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-config\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558406 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-cni-binary-copy\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558440 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a813e5e1-abad-433b-a1b3-24cab8119638-mcd-auth-proxy-config\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558469 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558569 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-ovn\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558645 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558686 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb4216ca-0206-4794-bb05-efabd0dc96cb-cni-binary-copy\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558715 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-cni-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558747 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-kubelet\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558766 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a813e5e1-abad-433b-a1b3-24cab8119638-rootfs\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558787 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-netd\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558822 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eb4216ca-0206-4794-bb05-efabd0dc96cb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558847 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-env-overrides\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558870 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ab5426-0159-4610-b378-f4c2e20b82b9-ovn-node-metrics-cert\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.558960 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dqjb\" (UniqueName: \"kubernetes.io/projected/a813e5e1-abad-433b-a1b3-24cab8119638-kube-api-access-8dqjb\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.559073 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-conf-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.574676 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.589036 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.601704 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.613625 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.630475 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.643639 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.658256 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659557 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-multus-certs\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659608 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-node-log\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659646 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-ovn-kubernetes\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659674 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-os-release\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659705 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-kubelet\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659734 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-netns\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659767 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdwn2\" (UniqueName: \"kubernetes.io/projected/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-kube-api-access-kdwn2\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659804 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-netns\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659840 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a813e5e1-abad-433b-a1b3-24cab8119638-proxy-tls\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659842 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-netns\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659873 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659899 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-os-release\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659937 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-netns\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659948 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-node-log\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660005 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-ovn-kubernetes\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659748 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-multus-certs\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.659947 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n84cc\" (UniqueName: \"kubernetes.io/projected/eb4216ca-0206-4794-bb05-efabd0dc96cb-kube-api-access-n84cc\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660043 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-kubelet\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660095 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-config\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660136 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-cni-binary-copy\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660159 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a813e5e1-abad-433b-a1b3-24cab8119638-mcd-auth-proxy-config\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660186 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb4216ca-0206-4794-bb05-efabd0dc96cb-cni-binary-copy\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660207 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-cni-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660227 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-kubelet\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660246 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660287 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-ovn\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660309 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660336 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a813e5e1-abad-433b-a1b3-24cab8119638-rootfs\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660360 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-kubelet\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660379 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eb4216ca-0206-4794-bb05-efabd0dc96cb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660516 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-netd\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660585 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-conf-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660610 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-env-overrides\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660613 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-cni-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660631 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ab5426-0159-4610-b378-f4c2e20b82b9-ovn-node-metrics-cert\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660679 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dqjb\" (UniqueName: \"kubernetes.io/projected/a813e5e1-abad-433b-a1b3-24cab8119638-kube-api-access-8dqjb\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660711 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660723 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-cnibin\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660754 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-systemd-units\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660782 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-os-release\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660785 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-ovn\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660787 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-cnibin\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660808 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-cnibin\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660816 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660835 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-socket-dir-parent\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660931 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-conf-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660933 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-netd\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660946 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a813e5e1-abad-433b-a1b3-24cab8119638-rootfs\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.660967 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-bin\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661025 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-etc-kubernetes\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661051 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxd2c\" (UniqueName: \"kubernetes.io/projected/41ab5426-0159-4610-b378-f4c2e20b82b9-kube-api-access-dxd2c\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661128 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-os-release\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661133 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-socket-dir-parent\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661150 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-systemd-units\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661172 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661209 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-cnibin\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661234 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-daemon-config\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661272 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-var-lib-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661295 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-log-socket\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661404 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-log-socket\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661456 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-k8s-cni-cncf-io\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661473 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eb4216ca-0206-4794-bb05-efabd0dc96cb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661498 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-bin\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661534 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-run-k8s-cni-cncf-io\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661566 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-system-cni-dir\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661595 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-system-cni-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661646 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-hostroot\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661675 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-systemd\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661703 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-etc-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661727 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-script-lib\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661770 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-system-cni-dir\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661791 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb4216ca-0206-4794-bb05-efabd0dc96cb-system-cni-dir\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-hostroot\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661846 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-systemd\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.661876 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-etc-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662020 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-multus-daemon-config\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662062 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-etc-kubernetes\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662089 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-var-lib-openvswitch\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662149 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-cni-bin\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662187 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-cni-multus\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662175 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb4216ca-0206-4794-bb05-efabd0dc96cb-cni-binary-copy\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662238 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-cni-multus\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662218 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-host-var-lib-cni-bin\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662279 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-slash\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662318 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-slash\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.662578 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-cni-binary-copy\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.664635 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-config\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.665674 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ab5426-0159-4610-b378-f4c2e20b82b9-ovn-node-metrics-cert\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.666492 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a813e5e1-abad-433b-a1b3-24cab8119638-mcd-auth-proxy-config\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.666563 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a813e5e1-abad-433b-a1b3-24cab8119638-proxy-tls\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.669896 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-env-overrides\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.670009 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-script-lib\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.678550 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.681620 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxd2c\" (UniqueName: \"kubernetes.io/projected/41ab5426-0159-4610-b378-f4c2e20b82b9-kube-api-access-dxd2c\") pod \"ovnkube-node-8dmqg\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.681668 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdwn2\" (UniqueName: \"kubernetes.io/projected/c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e-kube-api-access-kdwn2\") pod \"multus-2dqxx\" (UID: \"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\") " pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.687448 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n84cc\" (UniqueName: \"kubernetes.io/projected/eb4216ca-0206-4794-bb05-efabd0dc96cb-kube-api-access-n84cc\") pod \"multus-additional-cni-plugins-vm98j\" (UID: \"eb4216ca-0206-4794-bb05-efabd0dc96cb\") " pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.689100 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dqjb\" (UniqueName: \"kubernetes.io/projected/a813e5e1-abad-433b-a1b3-24cab8119638-kube-api-access-8dqjb\") pod \"machine-config-daemon-rfcbg\" (UID: \"a813e5e1-abad-433b-a1b3-24cab8119638\") " pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.697297 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.718717 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.733754 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.753667 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.766864 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.778562 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.779291 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.785704 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2dqxx" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.796080 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.796498 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.800904 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vm98j" Oct 02 07:16:42 crc kubenswrapper[4960]: W1002 07:16:42.801165 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda813e5e1_abad_433b_a1b3_24cab8119638.slice/crio-d302d4c94278632a689ff7b4bcce0abd316b33275439f29d62f918c68bdedd3f WatchSource:0}: Error finding container d302d4c94278632a689ff7b4bcce0abd316b33275439f29d62f918c68bdedd3f: Status 404 returned error can't find the container with id d302d4c94278632a689ff7b4bcce0abd316b33275439f29d62f918c68bdedd3f Oct 02 07:16:42 crc kubenswrapper[4960]: W1002 07:16:42.802529 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc96f3ba3_6aa3_4f51_a8cc_1b21ef16f29e.slice/crio-67c1a360441e2a2396e8ddab8b8e6fb0cfe55b67b7365e0b5d47b073a8e240fc WatchSource:0}: Error finding container 67c1a360441e2a2396e8ddab8b8e6fb0cfe55b67b7365e0b5d47b073a8e240fc: Status 404 returned error can't find the container with id 67c1a360441e2a2396e8ddab8b8e6fb0cfe55b67b7365e0b5d47b073a8e240fc Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.817061 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: W1002 07:16:42.826526 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41ab5426_0159_4610_b378_f4c2e20b82b9.slice/crio-2fe94077b19b160088ae5787181c9eb342d2e0beea62869eb4a1a253aeace771 WatchSource:0}: Error finding container 2fe94077b19b160088ae5787181c9eb342d2e0beea62869eb4a1a253aeace771: Status 404 returned error can't find the container with id 2fe94077b19b160088ae5787181c9eb342d2e0beea62869eb4a1a253aeace771 Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.835435 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:42 crc kubenswrapper[4960]: I1002 07:16:42.851163 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:42Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.268499 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.268639 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.268755 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:16:47.268719687 +0000 UTC m=+28.300665984 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.268797 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.268879 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:47.26886002 +0000 UTC m=+28.300806417 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.268992 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.269060 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.269098 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:47.269084796 +0000 UTC m=+28.301031083 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.329305 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.329475 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.370609 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.370677 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.370876 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.370884 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.370939 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.370900 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.370957 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.370986 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.371054 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:47.371037342 +0000 UTC m=+28.402983629 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:43 crc kubenswrapper[4960]: E1002 07:16:43.371209 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:47.371187626 +0000 UTC m=+28.403133923 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.504010 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb4216ca-0206-4794-bb05-efabd0dc96cb" containerID="f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9" exitCode=0 Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.504058 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerDied","Data":"f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.504685 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerStarted","Data":"5ee466b5afd88bf4867f0f9e64a1d135945877c0ea0b21408d8fd66d3ed56b3a"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.506635 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.508455 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741" exitCode=0 Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.508535 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.508561 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"2fe94077b19b160088ae5787181c9eb342d2e0beea62869eb4a1a253aeace771"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.512070 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerStarted","Data":"a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.512113 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerStarted","Data":"67c1a360441e2a2396e8ddab8b8e6fb0cfe55b67b7365e0b5d47b073a8e240fc"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.517332 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.517388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.517402 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"d302d4c94278632a689ff7b4bcce0abd316b33275439f29d62f918c68bdedd3f"} Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.534207 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.560635 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.583070 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.595643 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.609760 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.653108 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.671615 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.692299 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.733163 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.744998 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.757510 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.772651 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.791064 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.803754 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.831528 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.848732 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.873283 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.885697 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.897935 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.913800 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.934225 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.951163 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.953891 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-46hxp"] Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.954369 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.955619 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.956135 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.956135 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.956187 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.966684 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.979289 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:43 crc kubenswrapper[4960]: I1002 07:16:43.991577 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.004249 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.020854 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.035428 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.054106 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.071576 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.079104 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8573c7cf-f32d-4f64-a857-5136816b5173-host\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.079337 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8qkl\" (UniqueName: \"kubernetes.io/projected/8573c7cf-f32d-4f64-a857-5136816b5173-kube-api-access-c8qkl\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.079412 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8573c7cf-f32d-4f64-a857-5136816b5173-serviceca\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.093674 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.107296 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.122532 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.136240 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.150683 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.162366 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.178267 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.180684 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8qkl\" (UniqueName: \"kubernetes.io/projected/8573c7cf-f32d-4f64-a857-5136816b5173-kube-api-access-c8qkl\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.180728 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8573c7cf-f32d-4f64-a857-5136816b5173-serviceca\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.180772 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8573c7cf-f32d-4f64-a857-5136816b5173-host\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.180856 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8573c7cf-f32d-4f64-a857-5136816b5173-host\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.182138 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8573c7cf-f32d-4f64-a857-5136816b5173-serviceca\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.193909 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.201510 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8qkl\" (UniqueName: \"kubernetes.io/projected/8573c7cf-f32d-4f64-a857-5136816b5173-kube-api-access-c8qkl\") pod \"node-ca-46hxp\" (UID: \"8573c7cf-f32d-4f64-a857-5136816b5173\") " pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.209574 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.222903 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.329141 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:44 crc kubenswrapper[4960]: E1002 07:16:44.329719 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.330283 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:44 crc kubenswrapper[4960]: E1002 07:16:44.330353 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.490506 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-46hxp" Oct 02 07:16:44 crc kubenswrapper[4960]: W1002 07:16:44.505243 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8573c7cf_f32d_4f64_a857_5136816b5173.slice/crio-8f8e81fd8d1a22663b74fd97aa9327565906b41fb5ecbb1c4467f33ebebd5246 WatchSource:0}: Error finding container 8f8e81fd8d1a22663b74fd97aa9327565906b41fb5ecbb1c4467f33ebebd5246: Status 404 returned error can't find the container with id 8f8e81fd8d1a22663b74fd97aa9327565906b41fb5ecbb1c4467f33ebebd5246 Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.528547 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.528609 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.528620 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.529917 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-46hxp" event={"ID":"8573c7cf-f32d-4f64-a857-5136816b5173","Type":"ContainerStarted","Data":"8f8e81fd8d1a22663b74fd97aa9327565906b41fb5ecbb1c4467f33ebebd5246"} Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.532425 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb4216ca-0206-4794-bb05-efabd0dc96cb" containerID="3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0" exitCode=0 Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.532476 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerDied","Data":"3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0"} Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.552615 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.564963 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.576822 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.587814 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.597105 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.610295 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.627406 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.640686 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.663462 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.681760 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.696440 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.718562 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.743752 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:44 crc kubenswrapper[4960]: I1002 07:16:44.759765 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.329881 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:45 crc kubenswrapper[4960]: E1002 07:16:45.330538 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.541655 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.541737 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.541759 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.543932 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-46hxp" event={"ID":"8573c7cf-f32d-4f64-a857-5136816b5173","Type":"ContainerStarted","Data":"4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698"} Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.546908 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb4216ca-0206-4794-bb05-efabd0dc96cb" containerID="c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd" exitCode=0 Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.547025 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerDied","Data":"c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd"} Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.567093 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.594039 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.613556 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.641484 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.659301 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.679225 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.694358 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.716342 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.731685 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.747634 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.767220 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.778816 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.791882 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.804261 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.820397 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.834096 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.845423 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.861956 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.875048 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.887353 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.901367 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.916095 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.937618 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.953624 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.969704 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.978395 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.982266 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.982316 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.982331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.982449 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.985829 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:45Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.991548 4960 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.991920 4960 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.993442 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.993500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.993515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.993542 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:45 crc kubenswrapper[4960]: I1002 07:16:45.993557 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:45Z","lastTransitionTime":"2025-10-02T07:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.006508 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.025914 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.028338 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.031044 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.031138 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.031166 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.031207 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.031235 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.045859 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.051899 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.051970 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.052031 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.052061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.052085 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.071748 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.076863 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.076899 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.076911 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.076929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.076940 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.092381 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.096276 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.096355 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.096369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.096395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.096413 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.111431 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.111768 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.113690 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.113730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.113744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.113765 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.113779 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.216421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.216479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.216492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.216513 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.216529 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.320188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.320245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.320266 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.320292 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.320314 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.329838 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.330132 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.330197 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:46 crc kubenswrapper[4960]: E1002 07:16:46.330436 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.424234 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.424297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.424307 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.424326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.424336 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.527848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.527891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.527902 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.527917 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.527927 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.556205 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb4216ca-0206-4794-bb05-efabd0dc96cb" containerID="fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2" exitCode=0 Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.556292 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerDied","Data":"fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.588180 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.608342 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.623754 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.630168 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.630226 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.630244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.630272 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.630294 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.637729 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.658725 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.674470 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.689985 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.703671 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.719086 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.733469 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.735476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.735510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.735526 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.735554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.735568 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.748474 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.764862 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.782299 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.802730 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.838233 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.838298 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.838313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.838341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.838358 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.941712 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.941769 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.941782 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.941807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:46 crc kubenswrapper[4960]: I1002 07:16:46.941822 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:46Z","lastTransitionTime":"2025-10-02T07:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.044776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.044827 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.044840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.044860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.044871 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.148563 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.148614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.148625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.148645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.148655 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.156549 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.161819 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.166574 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.171469 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.187281 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.206786 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.224107 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.245184 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.251472 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.251540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.251560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.251586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.251614 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.265386 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.286093 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.299637 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.311948 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.312245 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.312414 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:16:55.312387538 +0000 UTC m=+36.344333825 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.312471 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.312533 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.312632 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.312697 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.312701 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:55.312679825 +0000 UTC m=+36.344626112 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.312761 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:55.312750207 +0000 UTC m=+36.344696704 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.324238 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.329246 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.329386 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.337643 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.352601 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.355498 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.355559 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.355576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.355603 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.355619 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.372511 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.387571 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.403304 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.413378 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.413437 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413634 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413657 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413696 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413704 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413754 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413770 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:55.41374446 +0000 UTC m=+36.445690757 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413773 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:47 crc kubenswrapper[4960]: E1002 07:16:47.413865 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:16:55.413836792 +0000 UTC m=+36.445783259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.417462 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.434696 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.456679 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.459079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.459136 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.459149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.459176 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.459191 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.470505 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.483415 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.500733 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.522364 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.540374 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.554106 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.565573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.566446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.566508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.566516 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb4216ca-0206-4794-bb05-efabd0dc96cb" containerID="b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f" exitCode=0 Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.566546 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.566646 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerDied","Data":"b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.566683 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.575527 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.576508 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.588540 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.602194 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.615151 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.627946 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.652274 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.664367 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.670491 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.670550 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.670568 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.670594 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.670615 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.678916 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.694418 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.707474 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.721441 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.735842 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.750309 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.763785 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.773069 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.773108 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.773117 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.773137 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.773151 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.780650 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.796479 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.821192 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.834261 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.849025 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.862207 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:47Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.875900 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.876038 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.876099 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.876160 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.876219 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.979191 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.979599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.979687 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.979779 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:47 crc kubenswrapper[4960]: I1002 07:16:47.979865 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:47Z","lastTransitionTime":"2025-10-02T07:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.083280 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.083326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.083338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.083358 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.083373 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.186800 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.187149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.187242 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.187384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.187464 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.290083 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.290135 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.290145 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.290162 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.290173 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.329115 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.329167 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:48 crc kubenswrapper[4960]: E1002 07:16:48.329305 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:48 crc kubenswrapper[4960]: E1002 07:16:48.329466 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.393604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.393651 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.393661 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.393679 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.393691 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.496651 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.496710 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.496725 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.496745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.496760 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.592421 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb4216ca-0206-4794-bb05-efabd0dc96cb" containerID="80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3" exitCode=0 Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.592485 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerDied","Data":"80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.599408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.599512 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.599530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.599552 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.599570 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.607219 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.624775 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.638150 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.651356 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.672994 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.688445 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.704674 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.704747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.704760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.704783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.704796 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.706629 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.732026 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.749969 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.793467 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.807697 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.807740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.807753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.807775 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.807788 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.838519 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.857281 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.867503 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.878346 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.892057 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:48Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.911319 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.911398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.911414 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.911437 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:48 crc kubenswrapper[4960]: I1002 07:16:48.911456 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:48Z","lastTransitionTime":"2025-10-02T07:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.015005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.015059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.015069 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.015127 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.015139 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.118450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.118502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.118513 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.118532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.118546 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.221766 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.221806 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.221819 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.221836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.221848 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.325320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.325371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.325384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.325410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.325427 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.329641 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:49 crc kubenswrapper[4960]: E1002 07:16:49.329780 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.427872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.427917 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.427929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.427953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.427968 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.531473 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.531521 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.531529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.531547 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.531559 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.601717 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.603446 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.603550 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.610873 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" event={"ID":"eb4216ca-0206-4794-bb05-efabd0dc96cb","Type":"ContainerStarted","Data":"f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.618918 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.631337 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.632016 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.632788 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.633767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.633834 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.633850 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.633872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.633883 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.646965 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.659933 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.674284 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.691140 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.713135 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.734161 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.737567 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.737614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.737624 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.737644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.737655 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.750134 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.762911 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.772454 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.785841 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.801422 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.819052 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.838894 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.840005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.840056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.840066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.840086 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.840097 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.853176 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.863652 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.878420 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.891325 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.907090 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.919528 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.932498 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.943050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.943096 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.943106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.943124 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.943138 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:49Z","lastTransitionTime":"2025-10-02T07:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.946207 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.962137 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.974175 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:49 crc kubenswrapper[4960]: I1002 07:16:49.989338 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.019142 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.040924 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.045916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.045993 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.046005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.046023 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.046040 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.062944 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.080761 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.149369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.149416 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.149427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.149448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.149463 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.252338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.252387 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.252398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.252415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.252427 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.329389 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.329492 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:50 crc kubenswrapper[4960]: E1002 07:16:50.329588 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:50 crc kubenswrapper[4960]: E1002 07:16:50.329657 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.345112 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.355121 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.355161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.355171 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.355188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.355199 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.365266 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.382230 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.402670 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.419957 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.435824 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.455847 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.457695 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.457744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.457760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.457781 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.457796 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.474277 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.500265 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.514585 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.530506 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.545628 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.560399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.560462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.560475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.560492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.560503 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.561298 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.577249 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.595349 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:50Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.614526 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.668787 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.668844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.668860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.668882 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.668895 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.771807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.771850 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.771858 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.771874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.771885 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.875034 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.875088 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.875098 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.875112 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.875124 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.981015 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.981062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.981073 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.981093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:50 crc kubenswrapper[4960]: I1002 07:16:50.981103 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:50Z","lastTransitionTime":"2025-10-02T07:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.084570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.084620 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.084629 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.084647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.084660 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.188092 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.188130 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.188140 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.188157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.188166 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.290656 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.290696 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.290709 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.290727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.290740 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.329453 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:51 crc kubenswrapper[4960]: E1002 07:16:51.329646 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.394059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.394114 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.394129 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.394177 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.394191 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.497356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.497399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.497412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.497431 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.497444 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.600146 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.600244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.600272 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.600307 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.600334 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.618031 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.703209 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.703324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.703349 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.703382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.703405 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.806544 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.806592 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.806608 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.806627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.806640 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.909021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.909064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.909074 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.909093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:51 crc kubenswrapper[4960]: I1002 07:16:51.909106 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:51Z","lastTransitionTime":"2025-10-02T07:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.040840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.040889 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.040909 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.040934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.040952 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.145074 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.145148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.145172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.145210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.145241 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.249151 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.249204 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.249214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.249232 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.249243 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.329137 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.329206 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:52 crc kubenswrapper[4960]: E1002 07:16:52.329324 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:52 crc kubenswrapper[4960]: E1002 07:16:52.329511 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.351630 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.351683 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.351694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.351754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.351781 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.462689 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.462727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.462740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.462761 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.462774 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.566378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.566462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.566477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.566500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.566512 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.625406 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/0.log" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.632450 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79" exitCode=1 Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.632551 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.633865 4960 scope.go:117] "RemoveContainer" containerID="28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.653230 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.671295 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.671346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.671358 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.671378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.671393 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.674636 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.691499 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.710191 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.734311 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"message\\\":\\\"1.881926 6281 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:16:51.882031 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:16:51.882092 6281 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:16:51.882122 6281 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:51.882147 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:16:51.882196 6281 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 07:16:51.882223 6281 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 07:16:51.882238 6281 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:16:51.882214 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:16:51.882264 6281 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:16:51.882296 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:51.882318 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:16:51.882375 6281 factory.go:656] Stopping watch factory\\\\nI1002 07:16:51.882337 6281 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:16:51.882442 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:16:51.882476 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.757387 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.776077 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.776136 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.776156 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.776184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.776206 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.779752 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.798460 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.814460 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.832198 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.851120 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.866114 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.879278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.879331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.879347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.879365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.879377 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.884162 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.899932 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.915261 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:52Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.982184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.982234 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.982252 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.982274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:52 crc kubenswrapper[4960]: I1002 07:16:52.982291 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:52Z","lastTransitionTime":"2025-10-02T07:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.084793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.084859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.084872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.084890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.084904 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.188243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.188294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.188309 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.188331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.188348 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.290671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.290709 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.290719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.290737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.290746 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.329468 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:53 crc kubenswrapper[4960]: E1002 07:16:53.329690 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.395160 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.395209 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.395223 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.395239 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.395251 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.498167 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.498211 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.498223 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.498241 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.498252 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.601750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.602064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.602075 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.602094 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.602106 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.639292 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/0.log" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.643037 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.643440 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.664043 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.682163 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.699221 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.705612 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.705685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.705721 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.705745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.705760 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.717360 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.744558 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.766019 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.782333 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.799334 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.808774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.808874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.808911 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.808934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.808950 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.825952 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"message\\\":\\\"1.881926 6281 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:16:51.882031 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:16:51.882092 6281 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:16:51.882122 6281 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:51.882147 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:16:51.882196 6281 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 07:16:51.882223 6281 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 07:16:51.882238 6281 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:16:51.882214 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:16:51.882264 6281 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:16:51.882296 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:51.882318 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:16:51.882375 6281 factory.go:656] Stopping watch factory\\\\nI1002 07:16:51.882337 6281 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:16:51.882442 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:16:51.882476 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.841856 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.858908 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.872097 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.886000 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.899082 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.910845 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:53Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.912197 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.912235 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.912246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.912263 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:53 crc kubenswrapper[4960]: I1002 07:16:53.912274 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:53Z","lastTransitionTime":"2025-10-02T07:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.015218 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.015259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.015273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.015290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.015304 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.119917 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.119995 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.120006 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.120026 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.120037 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.223124 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.223175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.223187 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.223208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.223219 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.326472 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.326520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.326532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.326553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.326568 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.329739 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:54 crc kubenswrapper[4960]: E1002 07:16:54.329863 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.330236 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:54 crc kubenswrapper[4960]: E1002 07:16:54.330375 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.434596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.434641 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.434654 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.434677 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.434691 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.537211 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.537257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.537271 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.537289 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.537301 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.640323 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.641311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.641359 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.641411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.641429 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.648785 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/1.log" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.649343 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/0.log" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.651907 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879" exitCode=1 Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.651960 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.652044 4960 scope.go:117] "RemoveContainer" containerID="28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.652914 4960 scope.go:117] "RemoveContainer" containerID="a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879" Oct 02 07:16:54 crc kubenswrapper[4960]: E1002 07:16:54.653139 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.668661 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.680708 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.698019 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.712380 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.731336 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"message\\\":\\\"1.881926 6281 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:16:51.882031 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:16:51.882092 6281 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:16:51.882122 6281 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:51.882147 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:16:51.882196 6281 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 07:16:51.882223 6281 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 07:16:51.882238 6281 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:16:51.882214 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:16:51.882264 6281 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:16:51.882296 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:51.882318 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:16:51.882375 6281 factory.go:656] Stopping watch factory\\\\nI1002 07:16:51.882337 6281 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:16:51.882442 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:16:51.882476 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.743666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.743706 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.743716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.743734 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.743745 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.749334 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.772785 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.788231 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.801999 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.815143 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.829436 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.841415 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.846547 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.846630 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.846646 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.846666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.846680 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.858107 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.873645 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.897937 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:54Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.950151 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.950203 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.950213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.950231 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:54 crc kubenswrapper[4960]: I1002 07:16:54.950242 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:54Z","lastTransitionTime":"2025-10-02T07:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.053303 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.053353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.053366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.053388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.053402 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.155854 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.155927 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.155936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.155956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.155990 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.223832 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t"] Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.224477 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.227176 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.227295 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.239351 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.254032 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.258996 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.259054 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.259066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.259086 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.259101 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.271138 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.284770 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.299721 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.312480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.312543 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9929w\" (UniqueName: \"kubernetes.io/projected/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-kube-api-access-9929w\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.312607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.312646 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.314408 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.329188 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.329345 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.329605 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.344925 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.359236 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.362002 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.362059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.362078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.362108 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.362127 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.374384 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.395835 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.413688 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.413814 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.413852 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9929w\" (UniqueName: \"kubernetes.io/projected/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-kube-api-access-9929w\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.413878 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.413846132 +0000 UTC m=+52.445792419 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.413925 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.413989 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.414019 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.414043 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.414077 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.414205 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.414266 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.414251063 +0000 UTC m=+52.446197350 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.414413 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.414563 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.414526199 +0000 UTC m=+52.446472516 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.416806 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.416863 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.417210 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.416888 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.420475 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.420601 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.42056583 +0000 UTC m=+52.452512117 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.423185 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.424869 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.434994 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9929w\" (UniqueName: \"kubernetes.io/projected/a6bf9ea2-e784-43e4-a08a-7bcd491c805a-kube-api-access-9929w\") pod \"ovnkube-control-plane-749d76644c-rr46t\" (UID: \"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.439842 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.455227 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.465029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.465097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.465120 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.465149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.465173 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.470157 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.492412 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"message\\\":\\\"1.881926 6281 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:16:51.882031 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:16:51.882092 6281 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:16:51.882122 6281 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:51.882147 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:16:51.882196 6281 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 07:16:51.882223 6281 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 07:16:51.882238 6281 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:16:51.882214 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:16:51.882264 6281 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:16:51.882296 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:51.882318 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:16:51.882375 6281 factory.go:656] Stopping watch factory\\\\nI1002 07:16:51.882337 6281 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:16:51.882442 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:16:51.882476 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:55Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.514834 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.515145 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.515195 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.515210 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:55 crc kubenswrapper[4960]: E1002 07:16:55.515285 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.515260975 +0000 UTC m=+52.547207262 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.538469 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" Oct 02 07:16:55 crc kubenswrapper[4960]: W1002 07:16:55.553727 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6bf9ea2_e784_43e4_a08a_7bcd491c805a.slice/crio-f63767eb9eb63c72983f32a1f811abafda75c0299f8701aff95850184944edff WatchSource:0}: Error finding container f63767eb9eb63c72983f32a1f811abafda75c0299f8701aff95850184944edff: Status 404 returned error can't find the container with id f63767eb9eb63c72983f32a1f811abafda75c0299f8701aff95850184944edff Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.569517 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.569580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.569589 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.569609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.569658 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.656249 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" event={"ID":"a6bf9ea2-e784-43e4-a08a-7bcd491c805a","Type":"ContainerStarted","Data":"f63767eb9eb63c72983f32a1f811abafda75c0299f8701aff95850184944edff"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.659789 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/1.log" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.672100 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.672166 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.672186 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.672218 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.672237 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.775537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.775626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.775647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.775676 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.775705 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.878888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.878943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.878956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.879030 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.879130 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.981820 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.981861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.981870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.981887 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:55 crc kubenswrapper[4960]: I1002 07:16:55.981897 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:55Z","lastTransitionTime":"2025-10-02T07:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.085130 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.085186 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.085196 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.085217 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.085229 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.168854 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.168925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.168943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.169005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.169027 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.189100 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.194647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.194716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.194733 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.194757 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.194770 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.209170 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.214190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.214225 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.214237 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.214255 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.214268 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.230077 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.236788 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.236840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.236851 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.236867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.236877 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.250738 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.255452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.255502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.255519 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.255543 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.255559 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.270099 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.270225 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.272131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.272178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.272191 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.272210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.272224 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.316805 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hhvkr"] Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.322045 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.322293 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.329336 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.329374 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.329502 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.329576 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.339658 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.353715 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.370394 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.375789 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.375844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.375859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.375879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.375893 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.386131 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.400216 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.416303 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.424205 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctnf7\" (UniqueName: \"kubernetes.io/projected/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-kube-api-access-ctnf7\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.424266 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.430264 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.452963 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.466587 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.478962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.479243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.479351 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.479426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.479484 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.483104 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.498187 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.517268 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"message\\\":\\\"1.881926 6281 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:16:51.882031 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:16:51.882092 6281 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:16:51.882122 6281 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:51.882147 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:16:51.882196 6281 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 07:16:51.882223 6281 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 07:16:51.882238 6281 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:16:51.882214 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:16:51.882264 6281 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:16:51.882296 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:51.882318 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:16:51.882375 6281 factory.go:656] Stopping watch factory\\\\nI1002 07:16:51.882337 6281 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:16:51.882442 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:16:51.882476 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.525707 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctnf7\" (UniqueName: \"kubernetes.io/projected/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-kube-api-access-ctnf7\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.525765 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.525903 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:16:56 crc kubenswrapper[4960]: E1002 07:16:56.526000 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:16:57.025966408 +0000 UTC m=+38.057912695 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.527956 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.540429 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.544694 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctnf7\" (UniqueName: \"kubernetes.io/projected/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-kube-api-access-ctnf7\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.559213 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.570989 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.581532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.581578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.581588 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.581606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.581616 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.581682 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.671140 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" event={"ID":"a6bf9ea2-e784-43e4-a08a-7bcd491c805a","Type":"ContainerStarted","Data":"8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.671211 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" event={"ID":"a6bf9ea2-e784-43e4-a08a-7bcd491c805a","Type":"ContainerStarted","Data":"164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.684248 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.684310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.684324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.684353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.684366 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.690398 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.704089 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.726738 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.745182 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.763927 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.778524 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.788798 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.789337 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.789360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.789387 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.789408 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.802293 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"message\\\":\\\"1.881926 6281 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:16:51.882031 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:16:51.882092 6281 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:16:51.882122 6281 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:51.882147 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:16:51.882196 6281 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 07:16:51.882223 6281 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 07:16:51.882238 6281 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:16:51.882214 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:16:51.882264 6281 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:16:51.882296 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:51.882318 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:16:51.882375 6281 factory.go:656] Stopping watch factory\\\\nI1002 07:16:51.882337 6281 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:16:51.882442 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:16:51.882476 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.813539 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.826592 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.837875 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.849227 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.862118 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.880744 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.892355 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.892422 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.892435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.892458 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.892469 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.899751 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.915140 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.930194 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.942961 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:16:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.995692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.995741 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.995750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.995769 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:56 crc kubenswrapper[4960]: I1002 07:16:56.995781 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:56Z","lastTransitionTime":"2025-10-02T07:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.030352 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:57 crc kubenswrapper[4960]: E1002 07:16:57.030543 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:16:57 crc kubenswrapper[4960]: E1002 07:16:57.030621 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:16:58.030597691 +0000 UTC m=+39.062543978 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.098845 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.098921 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.098938 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.098962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.098999 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.202164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.202243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.202254 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.202274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.202319 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.305532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.305583 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.305594 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.305617 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.305629 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.329870 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:57 crc kubenswrapper[4960]: E1002 07:16:57.330277 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.409444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.409528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.409540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.409559 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.409572 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.512570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.512636 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.512649 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.512674 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.512688 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.615480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.615515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.615527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.615544 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.615554 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.719066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.719107 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.719121 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.719141 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.719154 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.822712 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.822808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.822826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.822851 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.822866 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.925771 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.926183 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.926275 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.926378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:57 crc kubenswrapper[4960]: I1002 07:16:57.926475 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:57Z","lastTransitionTime":"2025-10-02T07:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.029761 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.030131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.030291 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.030410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.030503 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.041366 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:58 crc kubenswrapper[4960]: E1002 07:16:58.041685 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:16:58 crc kubenswrapper[4960]: E1002 07:16:58.041943 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:17:00.041913438 +0000 UTC m=+41.073859725 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.133817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.134146 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.134223 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.134294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.134354 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.237665 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.237735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.237746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.237765 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.237779 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.329652 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.329698 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:16:58 crc kubenswrapper[4960]: E1002 07:16:58.329819 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.329876 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:16:58 crc kubenswrapper[4960]: E1002 07:16:58.330018 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:16:58 crc kubenswrapper[4960]: E1002 07:16:58.330104 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.344393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.344427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.344435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.344450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.344462 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.447385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.447435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.447446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.447464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.447475 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.549388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.549425 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.549446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.549461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.549471 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.652180 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.652239 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.652257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.652279 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.652298 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.756441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.756558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.756583 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.756621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.756657 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.859694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.859733 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.859743 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.859761 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.859772 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.962294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.962345 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.962356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.962373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:58 crc kubenswrapper[4960]: I1002 07:16:58.962385 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:58Z","lastTransitionTime":"2025-10-02T07:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.065722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.065791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.065803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.065827 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.065842 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.168918 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.168998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.169021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.169042 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.169062 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.271962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.272019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.272029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.272046 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.272058 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.329094 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:16:59 crc kubenswrapper[4960]: E1002 07:16:59.329225 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.374321 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.374371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.374383 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.374404 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.374419 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.477008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.477053 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.477067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.477090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.477102 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.579871 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.579924 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.579936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.579957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.579988 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.682101 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.682146 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.682155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.682170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.682180 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.784780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.784816 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.784825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.784841 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.784852 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.887520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.887554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.887564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.887579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.887589 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.990240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.990289 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.990298 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.990312 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:16:59 crc kubenswrapper[4960]: I1002 07:16:59.990324 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:16:59Z","lastTransitionTime":"2025-10-02T07:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.064836 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:00 crc kubenswrapper[4960]: E1002 07:17:00.065208 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:00 crc kubenswrapper[4960]: E1002 07:17:00.065361 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:17:04.065328894 +0000 UTC m=+45.097275181 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.093239 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.093306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.093330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.093365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.093389 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.196230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.196275 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.196285 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.196304 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.196315 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.299234 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.299290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.299307 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.299333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.299351 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.329505 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.329645 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.329847 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:00 crc kubenswrapper[4960]: E1002 07:17:00.329867 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:00 crc kubenswrapper[4960]: E1002 07:17:00.330040 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:00 crc kubenswrapper[4960]: E1002 07:17:00.330103 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.342999 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.354502 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.369336 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.383862 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.397377 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.401657 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.401732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.401741 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.401759 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.401768 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.413262 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.435799 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28849877860ddd016702592fe690ec9a75969054feb4d5a628bc330402734c79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"message\\\":\\\"1.881926 6281 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:16:51.882031 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:16:51.882092 6281 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:16:51.882122 6281 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:51.882147 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:16:51.882196 6281 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 07:16:51.882223 6281 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 07:16:51.882238 6281 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:16:51.882214 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:16:51.882264 6281 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:16:51.882296 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:51.882318 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:16:51.882375 6281 factory.go:656] Stopping watch factory\\\\nI1002 07:16:51.882337 6281 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:16:51.882442 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:16:51.882476 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.452667 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.468127 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.490746 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.505173 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.505229 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.505242 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.505262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.505278 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.506106 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.520927 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.534216 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.549700 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.563008 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.575927 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.590289 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.607388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.607449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.607458 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.607475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.607484 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.710065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.710109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.710119 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.710142 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.710154 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.813389 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.814045 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.814072 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.814090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.814102 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.916558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.916894 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.916999 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.917079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:00 crc kubenswrapper[4960]: I1002 07:17:00.917155 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:00Z","lastTransitionTime":"2025-10-02T07:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.020258 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.020294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.020305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.020320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.020330 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.123427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.123491 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.123511 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.123539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.123557 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.226774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.227311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.227423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.227576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.227670 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.329538 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:01 crc kubenswrapper[4960]: E1002 07:17:01.329720 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.331686 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.331925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.331938 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.332055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.332143 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.435774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.436115 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.436177 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.436245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.436314 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.539482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.539536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.539547 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.539568 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.539580 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.643024 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.643082 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.643093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.643113 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.643456 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.746729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.746817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.746842 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.746877 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.746901 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.849903 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.849959 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.849971 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.850004 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.850013 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.955127 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.955741 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.956308 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.956752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:01 crc kubenswrapper[4960]: I1002 07:17:01.956930 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:01Z","lastTransitionTime":"2025-10-02T07:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.060738 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.060804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.060815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.060835 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.060847 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.164063 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.164116 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.164126 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.164145 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.164158 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.267649 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.267726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.267736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.267754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.267766 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.329610 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.329725 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:02 crc kubenswrapper[4960]: E1002 07:17:02.329761 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.329838 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:02 crc kubenswrapper[4960]: E1002 07:17:02.329906 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:02 crc kubenswrapper[4960]: E1002 07:17:02.330047 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.370573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.370618 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.370626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.370645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.370655 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.472900 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.472943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.472961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.473008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.473022 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.576434 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.576477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.576486 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.576509 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.576522 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.679062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.679138 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.679170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.679193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.679207 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.781399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.781454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.781464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.781481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.781491 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.884227 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.884265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.884275 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.884294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.884306 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.986708 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.986754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.986769 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.986810 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:02 crc kubenswrapper[4960]: I1002 07:17:02.986823 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:02Z","lastTransitionTime":"2025-10-02T07:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.089830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.089869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.089877 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.089895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.089904 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.192481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.192519 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.192527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.192544 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.192554 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.295486 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.295566 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.295584 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.295614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.295632 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.329563 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:03 crc kubenswrapper[4960]: E1002 07:17:03.329804 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.398143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.398206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.398219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.398240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.398251 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.501497 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.501551 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.501560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.501582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.501599 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.604012 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.604048 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.604057 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.604073 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.604082 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.705963 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.706022 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.706032 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.706050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.706060 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.808762 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.808803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.808815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.808834 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.808846 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.912021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.912066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.912076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.912104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:03 crc kubenswrapper[4960]: I1002 07:17:03.912118 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:03Z","lastTransitionTime":"2025-10-02T07:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.019523 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.019577 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.019587 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.019630 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.019735 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.109030 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:04 crc kubenswrapper[4960]: E1002 07:17:04.109268 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:04 crc kubenswrapper[4960]: E1002 07:17:04.109395 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:17:12.109364274 +0000 UTC m=+53.141310561 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.123471 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.123529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.123537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.123552 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.123580 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.226295 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.226607 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.226679 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.226752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.226826 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.329129 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.329129 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.329253 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:04 crc kubenswrapper[4960]: E1002 07:17:04.329872 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:04 crc kubenswrapper[4960]: E1002 07:17:04.329943 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.329533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.329994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.330012 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.330029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.330043 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: E1002 07:17:04.330333 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.432576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.432631 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.432667 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.432687 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.432700 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.535905 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.535993 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.536008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.536031 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.536044 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.639039 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.639098 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.639108 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.639126 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.639141 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.742114 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.742256 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.742281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.742311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.742334 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.846031 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.846093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.846105 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.846131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.846144 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.949220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.949269 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.949280 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.949297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:04 crc kubenswrapper[4960]: I1002 07:17:04.949306 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:04Z","lastTransitionTime":"2025-10-02T07:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.052634 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.052688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.052698 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.052715 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.052726 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.156172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.156245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.156262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.156292 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.156310 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.259809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.259856 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.259868 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.259889 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.259902 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.329125 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:05 crc kubenswrapper[4960]: E1002 07:17:05.329275 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.362701 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.362745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.362758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.362783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.362799 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.465958 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.466038 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.466053 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.466080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.466097 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.568911 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.568965 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.568997 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.569020 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.569034 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.672281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.672352 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.672364 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.672384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.672399 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.775228 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.775289 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.775302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.775322 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.775333 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.784529 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.786049 4960 scope.go:117] "RemoveContainer" containerID="a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.804005 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.818844 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.831836 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.843866 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.857509 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.877510 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.878369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.878411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.878424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.878442 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.878456 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.896648 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.914937 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.930259 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.946258 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.967093 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.982923 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.983002 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.983019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.983041 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:05 crc kubenswrapper[4960]: I1002 07:17:05.983050 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:05Z","lastTransitionTime":"2025-10-02T07:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.003047 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.035652 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.055295 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.076084 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.085855 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.085895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.085906 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.085922 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.085934 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.092857 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.107051 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.189265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.189328 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.189341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.189376 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.189394 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.293493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.293537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.293545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.293563 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.293572 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.329623 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.329623 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.329682 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.329807 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.329940 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.330079 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.397002 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.397045 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.397056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.397077 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.397089 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.487265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.487346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.487360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.487385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.487399 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.503799 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.508808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.508871 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.508884 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.509000 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.509019 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.525628 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.531848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.531904 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.531922 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.531956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.532004 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.546096 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.550522 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.550566 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.550577 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.550596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.550625 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.561847 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.566064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.566096 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.566106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.566119 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.566129 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.579864 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.579993 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.581750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.581798 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.581809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.581831 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.581843 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.685333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.685776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.685789 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.685809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.685826 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.708936 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/2.log" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.709710 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/1.log" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.713308 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92" exitCode=1 Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.713362 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.713412 4960 scope.go:117] "RemoveContainer" containerID="a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.714740 4960 scope.go:117] "RemoveContainer" containerID="778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92" Oct 02 07:17:06 crc kubenswrapper[4960]: E1002 07:17:06.715251 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.733749 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.750165 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.765566 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.789575 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.789621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.789633 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.789652 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.789666 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.802294 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.818707 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.830936 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.851522 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.865958 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.886049 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.892052 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.892113 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.892123 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.892143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.892154 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.901209 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.915287 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.926191 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.941390 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.954396 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.972084 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.991026 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.995085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.995138 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.995148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.995169 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:06 crc kubenswrapper[4960]: I1002 07:17:06.995178 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:06Z","lastTransitionTime":"2025-10-02T07:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.006008 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.097749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.097794 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.097804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.097820 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.097832 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.200757 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.200802 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.200815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.200832 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.200843 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.304331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.304368 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.304378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.304395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.304409 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.329147 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:07 crc kubenswrapper[4960]: E1002 07:17:07.329295 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.407997 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.408048 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.408058 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.408076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.408090 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.512090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.512160 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.512169 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.512188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.512202 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.615685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.615743 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.615755 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.615781 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.615795 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.719003 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.719053 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.719072 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.719396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.719433 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.721671 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/2.log" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.824116 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.824208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.824262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.824299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.824337 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.928011 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.928106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.928133 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.928170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:07 crc kubenswrapper[4960]: I1002 07:17:07.928242 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:07Z","lastTransitionTime":"2025-10-02T07:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.031120 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.031181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.031195 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.031219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.031235 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.134341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.134406 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.134427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.134453 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.134472 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.238153 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.238236 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.238260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.238294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.238319 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.329737 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.329873 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:08 crc kubenswrapper[4960]: E1002 07:17:08.330026 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:08 crc kubenswrapper[4960]: E1002 07:17:08.330206 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.330522 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:08 crc kubenswrapper[4960]: E1002 07:17:08.330675 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.342193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.342289 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.342313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.342343 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.342366 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.446352 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.446410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.446426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.446456 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.446473 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.551299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.551388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.551405 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.551441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.551459 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.654628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.654678 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.654688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.654705 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.654717 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.758055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.758115 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.758126 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.758149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.758163 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.861878 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.861947 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.861966 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.862028 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.862049 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.964510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.964593 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.964611 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.964645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:08 crc kubenswrapper[4960]: I1002 07:17:08.964665 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:08Z","lastTransitionTime":"2025-10-02T07:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.068766 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.068825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.068842 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.068870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.068890 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.171669 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.171736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.171754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.171789 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.171810 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.275281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.275350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.275365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.275393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.275412 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.329478 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:09 crc kubenswrapper[4960]: E1002 07:17:09.329634 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.378253 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.378339 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.378348 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.378395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.378408 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.482131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.482182 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.482196 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.482218 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.482231 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.584691 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.584744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.584758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.584780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.584793 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.687208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.687247 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.687255 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.687272 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.687282 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.789501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.789563 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.789581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.789606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.789623 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.892098 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.892143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.892155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.892176 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.892190 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.994852 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.994907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.994918 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.994937 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:09 crc kubenswrapper[4960]: I1002 07:17:09.994949 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:09Z","lastTransitionTime":"2025-10-02T07:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.098043 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.098098 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.098109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.098128 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.098139 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.200460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.200506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.200518 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.200538 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.200553 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.302959 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.303037 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.303047 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.303064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.303077 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.329225 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.329225 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.329351 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:10 crc kubenswrapper[4960]: E1002 07:17:10.329439 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:10 crc kubenswrapper[4960]: E1002 07:17:10.329862 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:10 crc kubenswrapper[4960]: E1002 07:17:10.329932 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.345963 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.365668 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.391097 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.407116 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.407574 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.407740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.407875 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.408044 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.409264 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.430003 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.447003 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.461600 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.490218 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.510537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.510589 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.510600 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.510619 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.510632 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.515620 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.529785 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.548085 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.563247 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.582826 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.597253 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.612616 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.613746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.613799 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.613813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.613837 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.613851 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.625192 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.642812 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.716965 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.717059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.717074 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.717098 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.717111 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.820419 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.820463 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.820474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.820492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.820503 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.922864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.922901 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.922910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.922926 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:10 crc kubenswrapper[4960]: I1002 07:17:10.922937 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:10Z","lastTransitionTime":"2025-10-02T07:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.025194 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.025237 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.025246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.025263 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.025274 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.128654 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.128702 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.128711 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.128728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.128738 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.232641 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.232671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.232681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.232696 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.232724 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.329779 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.329962 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.341479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.341524 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.341537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.341553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.341563 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.444457 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.444516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.444528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.444546 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.444559 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.503068 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503222 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:43.503188517 +0000 UTC m=+84.535134804 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.503361 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503554 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503597 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:43.503590256 +0000 UTC m=+84.535536543 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.503550 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.503656 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503709 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503743 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:43.5037366 +0000 UTC m=+84.535682887 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503784 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503821 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503844 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.503951 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:43.503923564 +0000 UTC m=+84.535869891 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.547740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.547812 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.547823 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.547844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.547856 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.604960 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.605287 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.605355 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.605372 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4960]: E1002 07:17:11.605469 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:43.60543616 +0000 UTC m=+84.637382647 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.651196 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.651261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.651278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.651300 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.651316 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.754138 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.754492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.754568 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.754645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.754718 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.858026 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.858101 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.858120 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.858157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.858177 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.961500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.961626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.961647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.961679 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:11 crc kubenswrapper[4960]: I1002 07:17:11.961705 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:11Z","lastTransitionTime":"2025-10-02T07:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.066530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.066610 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.066630 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.066656 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.066679 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.111649 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:12 crc kubenswrapper[4960]: E1002 07:17:12.111829 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:12 crc kubenswrapper[4960]: E1002 07:17:12.111904 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:17:28.111880027 +0000 UTC m=+69.143826314 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.172151 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.174891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.174928 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.174953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.174965 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.278730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.278803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.278822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.278854 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.278872 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.329649 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.329802 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:12 crc kubenswrapper[4960]: E1002 07:17:12.329911 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:12 crc kubenswrapper[4960]: E1002 07:17:12.330041 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.330365 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:12 crc kubenswrapper[4960]: E1002 07:17:12.330550 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.382262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.382304 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.382320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.382340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.382355 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.485782 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.485834 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.485844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.485861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.485872 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.588691 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.588813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.588839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.588872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.588900 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.692412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.692465 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.692477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.692499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.692511 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.794759 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.794805 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.794824 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.794848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.794865 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.898531 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.898599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.898618 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.898647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:12 crc kubenswrapper[4960]: I1002 07:17:12.898665 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:12Z","lastTransitionTime":"2025-10-02T07:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.002521 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.002570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.002580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.002598 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.002613 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.118160 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.118362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.118386 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.118415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.118432 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.223210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.223273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.223296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.223327 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.223352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.326461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.326518 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.326531 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.326553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.326566 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.329347 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:13 crc kubenswrapper[4960]: E1002 07:17:13.329594 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.432685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.433142 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.433318 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.433454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.433586 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.536431 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.536463 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.536472 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.536485 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.536493 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.638519 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.638793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.638883 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.639027 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.639115 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.741908 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.742206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.742279 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.742341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.742406 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.844432 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.844511 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.844533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.844573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.844599 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.947017 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.947322 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.947483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.947590 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:13 crc kubenswrapper[4960]: I1002 07:17:13.947696 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:13Z","lastTransitionTime":"2025-10-02T07:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.055950 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.056247 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.056334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.056452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.056543 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.159698 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.159745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.159758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.159779 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.159789 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.262128 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.262193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.262204 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.262222 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.262235 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.329249 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.329395 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:14 crc kubenswrapper[4960]: E1002 07:17:14.329496 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.329587 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:14 crc kubenswrapper[4960]: E1002 07:17:14.329718 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:14 crc kubenswrapper[4960]: E1002 07:17:14.329871 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.365286 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.365349 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.365363 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.365385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.365397 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.468554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.468606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.468661 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.468684 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.468702 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.571308 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.571362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.571380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.571404 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.571425 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.674066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.674131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.674150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.674175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.674193 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.776922 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.776964 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.776987 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.777003 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.777013 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.879657 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.880071 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.880253 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.880407 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.880556 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.984736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.985181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.985315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.985461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:14 crc kubenswrapper[4960]: I1002 07:17:14.985591 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:14Z","lastTransitionTime":"2025-10-02T07:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.089253 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.089317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.089328 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.089350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.089363 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.192491 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.192526 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.192537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.192554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.192563 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.294435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.294485 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.294494 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.294510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.294527 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.329204 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:15 crc kubenswrapper[4960]: E1002 07:17:15.329337 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.397757 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.397791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.397800 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.397817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.397826 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.500572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.500629 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.500640 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.500654 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.500851 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.603994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.604053 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.604063 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.604081 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.604094 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.706489 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.706527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.706536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.706550 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.706560 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.809558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.809605 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.809614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.809642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.809653 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.912489 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.912554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.912565 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.912582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4960]: I1002 07:17:15.912593 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.016064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.016127 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.016143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.016167 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.016184 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.123430 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.123475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.123486 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.123502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.123512 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.226580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.226642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.226661 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.226686 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.226702 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.328913 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.329020 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.328947 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.329165 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.329319 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.329425 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.329679 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.329709 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.329721 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.329740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.329752 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.432574 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.432626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.432642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.432692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.432704 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.535450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.535495 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.535720 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.535764 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.535777 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.639592 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.639680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.639747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.639769 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.639781 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.643317 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.655853 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.666748 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.687847 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.706944 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.727274 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.744437 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.744492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.744505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.744523 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.744816 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.753816 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.753897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.753910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.753957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.754382 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.758827 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.777819 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.782202 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.782229 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.782238 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.782252 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.782261 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.793341 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.799347 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.803869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.804013 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.804119 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.804210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.804312 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.807343 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.823692 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.823666 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.827822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.827859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.827872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.827889 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.827904 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.838325 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.842674 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.846574 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.846621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.846632 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.846650 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.846663 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.857844 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: E1002 07:17:16.858025 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.860748 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.860791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.860804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.860824 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.860837 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.864322 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.877991 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.888382 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.904686 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.931323 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.941275 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.952347 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.962531 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.963446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.963475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.963483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.963500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4960]: I1002 07:17:16.963510 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.066267 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.066301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.066309 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.066325 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.066335 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.168849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.168907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.168917 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.168931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.168940 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.271272 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.271308 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.271361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.271378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.271400 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.329146 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:17 crc kubenswrapper[4960]: E1002 07:17:17.329290 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.373944 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.373990 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.374000 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.374016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.374025 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.476398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.476446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.476458 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.476476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.476489 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.578133 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.578170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.578181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.578199 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.578210 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.680051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.680084 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.680093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.680107 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.680115 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.782412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.782441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.782454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.782469 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.782480 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.884668 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.884716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.884725 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.884742 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.884754 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.986840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.986890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.986901 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.986922 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4960]: I1002 07:17:17.986934 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.089204 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.089244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.089256 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.089275 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.089287 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.192239 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.192293 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.192302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.192322 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.192333 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.294751 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.294787 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.294795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.294808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.294817 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.329490 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.329661 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:18 crc kubenswrapper[4960]: E1002 07:17:18.329751 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.329810 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:18 crc kubenswrapper[4960]: E1002 07:17:18.330046 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:18 crc kubenswrapper[4960]: E1002 07:17:18.330097 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.397310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.397551 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.397609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.397708 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.397768 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.499832 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.500061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.500077 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.500095 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.500104 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.602496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.602538 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.602549 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.602565 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.602578 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.705277 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.705314 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.705326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.705342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.705353 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.807009 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.807043 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.807052 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.807065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.807076 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.909083 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.909115 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.909124 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.909141 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4960]: I1002 07:17:18.909153 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.011362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.011385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.011393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.011409 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.011441 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.114021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.114066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.114081 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.114099 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.114111 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.217276 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.217321 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.217331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.217347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.217357 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.319274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.319314 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.319326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.319348 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.319359 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.329261 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:19 crc kubenswrapper[4960]: E1002 07:17:19.329430 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.422945 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.423536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.423618 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.423692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.423783 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.526580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.526641 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.526655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.526678 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.526691 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.629616 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.629913 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.629995 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.630079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.630152 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.732839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.733153 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.733232 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.733316 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.733380 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.835374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.835821 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.835839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.835860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.835873 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.938847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.938882 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.938891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.938905 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4960]: I1002 07:17:19.938915 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.040966 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.041269 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.041347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.041421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.041495 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.143497 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.143567 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.143579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.143596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.143633 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.246281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.246325 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.246336 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.246353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.246364 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.329590 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:20 crc kubenswrapper[4960]: E1002 07:17:20.330792 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.330092 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:20 crc kubenswrapper[4960]: E1002 07:17:20.332826 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.330064 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:20 crc kubenswrapper[4960]: E1002 07:17:20.333050 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.331449 4960 scope.go:117] "RemoveContainer" containerID="778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92" Oct 02 07:17:20 crc kubenswrapper[4960]: E1002 07:17:20.333341 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.348131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.348162 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.348171 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.348186 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.348198 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.354948 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.372579 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.386135 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.399706 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.410404 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.421775 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.433208 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.445045 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.450825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.450867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.450879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.450901 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.450912 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.457080 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.468569 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.480666 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.499720 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.514347 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.532121 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.544833 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.553260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.553290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.553298 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.553313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.553324 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.559477 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.571010 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.588859 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a66841a07ab2d6074543fc8e21f64f2e759d96baed9f88108bf91ac40e6e0879\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651685 6401 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:16:53.651780 6401 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:16:53.652144 6401 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:16:53.652183 6401 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:16:53.652232 6401 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:16:53.652283 6401 factory.go:656] Stopping watch factory\\\\nI1002 07:16:53.652304 6401 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:16:53.662484 6401 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 07:16:53.662506 6401 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 07:16:53.662591 6401 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:16:53.662615 6401 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:16:53.662699 6401 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.601052 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.612319 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.624615 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.642418 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.655911 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.655935 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.655943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.655956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.655965 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.656967 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.667573 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.692053 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.704654 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.717404 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.727114 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.738223 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.749105 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.759516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.759554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.759564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.759581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.759593 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.760928 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.775720 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.786864 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.799201 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.810672 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.822167 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.862430 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.862470 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.862479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.862495 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.862507 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.965017 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.965068 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.965078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.965096 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4960]: I1002 07:17:20.965112 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.068332 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.068377 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.068393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.068415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.068429 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.171234 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.171268 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.171278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.171293 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.171302 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.273313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.273356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.273366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.273381 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.273391 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.329662 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:21 crc kubenswrapper[4960]: E1002 07:17:21.329889 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.376097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.376140 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.376161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.376181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.376191 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.479398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.480076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.480149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.480233 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.480293 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.583159 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.583497 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.583570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.583656 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.583740 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.687113 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.687161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.687172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.687188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.687199 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.793508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.794255 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.794295 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.794330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.794365 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.897791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.897906 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.898351 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.898382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4960]: I1002 07:17:21.898399 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.000934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.001028 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.001047 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.001073 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.001094 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.103161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.103200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.103208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.103223 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.103233 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.205877 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.205916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.205925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.205938 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.205947 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.308939 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.308993 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.309005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.309021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.309030 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.329724 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:22 crc kubenswrapper[4960]: E1002 07:17:22.329840 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.329732 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.329732 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:22 crc kubenswrapper[4960]: E1002 07:17:22.329904 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:22 crc kubenswrapper[4960]: E1002 07:17:22.330087 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.412125 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.412170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.412182 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.412198 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.412210 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.514937 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.516200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.516430 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.516592 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.516739 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.619936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.620009 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.620021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.620047 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.620078 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.722803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.722884 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.722902 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.722935 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.722954 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.827206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.827265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.827282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.827312 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.827324 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.930940 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.931007 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.931021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.931041 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4960]: I1002 07:17:22.931065 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.033254 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.033290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.033301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.033316 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.033326 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.136315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.136427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.136458 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.136512 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.136535 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.239581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.239628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.239645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.239664 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.239680 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.329315 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:23 crc kubenswrapper[4960]: E1002 07:17:23.329446 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.342493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.342531 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.342619 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.342658 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.342669 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.446995 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.447060 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.447072 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.447091 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.447103 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.550100 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.550174 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.550195 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.550230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.550251 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.652531 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.652606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.652623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.652652 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.652679 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.755830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.755869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.755878 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.755895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.755904 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.858368 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.858454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.858477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.858504 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.858523 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.961756 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.961816 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.961833 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.961859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4960]: I1002 07:17:23.961879 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.065304 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.065342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.065351 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.065371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.065382 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.168802 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.168853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.168865 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.168887 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.168897 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.273000 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.273055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.273064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.273104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.273113 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.329191 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.329677 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.330115 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:24 crc kubenswrapper[4960]: E1002 07:17:24.330093 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:24 crc kubenswrapper[4960]: E1002 07:17:24.330292 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:24 crc kubenswrapper[4960]: E1002 07:17:24.330487 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.376141 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.376190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.376202 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.376220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.376235 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.479118 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.479173 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.479190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.479215 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.479252 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.581817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.581886 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.581909 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.581942 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.581962 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.684449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.684488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.684500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.684515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.684528 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.786496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.787061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.787264 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.787461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.787656 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.890629 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.890675 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.890685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.890705 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.890716 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.994179 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.994246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.994270 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.994302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4960]: I1002 07:17:24.994324 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.098730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.098790 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.098804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.098825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.098839 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.201600 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.201662 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.201680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.201705 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.201723 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.305502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.305562 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.305590 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.305621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.305638 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.329111 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:25 crc kubenswrapper[4960]: E1002 07:17:25.329275 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.408808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.408903 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.408930 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.408965 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.409027 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.512587 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.512667 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.512692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.512727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.512750 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.615907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.615953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.615964 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.616004 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.616017 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.718767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.718813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.718825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.718844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.718858 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.821572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.821866 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.821881 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.821938 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.821952 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.925349 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.925413 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.925424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.925482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4960]: I1002 07:17:25.925492 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.028065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.028143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.028161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.028189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.028218 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.131660 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.131710 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.131726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.131746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.131759 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.235374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.235423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.235436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.235453 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.235465 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.329867 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.329918 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.329927 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:26 crc kubenswrapper[4960]: E1002 07:17:26.330136 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:26 crc kubenswrapper[4960]: E1002 07:17:26.330217 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:26 crc kubenswrapper[4960]: E1002 07:17:26.330354 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.337216 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.337257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.337268 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.337287 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.337296 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.439921 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.439964 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.439994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.440015 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.440027 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.543290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.543347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.543361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.543385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.543405 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.647140 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.647189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.647200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.647221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.647233 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.750435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.750490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.750500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.750518 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.750529 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.855059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.855137 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.855157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.855189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.855212 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.958753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.958817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.958831 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.958895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4960]: I1002 07:17:26.958924 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.060778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.060828 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.060840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.060856 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.060867 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.163713 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.163770 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.163785 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.163806 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.163821 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.248369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.248405 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.248413 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.248427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.248436 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: E1002 07:17:27.268054 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.273316 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.273347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.273358 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.273377 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.273388 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: E1002 07:17:27.289533 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.294643 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.294670 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.294679 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.294694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.294703 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: E1002 07:17:27.312258 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.316859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.316903 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.316916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.316940 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.316956 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.328773 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:27 crc kubenswrapper[4960]: E1002 07:17:27.328905 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:27 crc kubenswrapper[4960]: E1002 07:17:27.331431 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.336959 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.337040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.337055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.337076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.337090 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: E1002 07:17:27.356158 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4960]: E1002 07:17:27.356328 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.358654 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.358709 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.358727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.358750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.358769 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.462464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.462506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.462515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.462534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.462550 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.565577 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.565629 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.565642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.565662 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.565673 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.668825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.668884 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.668902 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.668929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.668948 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.772069 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.772145 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.772161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.772183 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.772196 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.875693 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.875810 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.875832 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.875864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.875883 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.979584 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.979683 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.979704 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.979742 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4960]: I1002 07:17:27.979763 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.083557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.083599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.083609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.083625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.083636 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.187652 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.187703 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.187717 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.187743 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.187760 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.197040 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:28 crc kubenswrapper[4960]: E1002 07:17:28.197214 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:28 crc kubenswrapper[4960]: E1002 07:17:28.197293 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:18:00.197270528 +0000 UTC m=+101.229216815 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.291209 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.291273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.291295 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.291339 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.291375 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.329050 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.329088 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.329272 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:28 crc kubenswrapper[4960]: E1002 07:17:28.329261 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:28 crc kubenswrapper[4960]: E1002 07:17:28.329415 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:28 crc kubenswrapper[4960]: E1002 07:17:28.329494 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.396209 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.396319 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.396343 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.396375 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.396405 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.499829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.499897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.499912 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.499939 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.499958 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.603172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.603247 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.603264 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.603295 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.603314 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.705285 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.705347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.705358 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.705373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.705385 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.808498 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.808564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.808580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.808610 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.808627 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.912334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.912390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.912401 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.912422 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4960]: I1002 07:17:28.912433 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.016208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.016282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.016299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.016321 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.016337 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.119669 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.119722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.119732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.119756 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.119767 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.222419 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.222515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.222539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.222574 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.222598 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.325333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.325379 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.325390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.325410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.325422 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.329205 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:29 crc kubenswrapper[4960]: E1002 07:17:29.329416 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.428398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.428443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.428474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.428494 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.428506 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.531418 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.531478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.531489 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.531508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.531520 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.635443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.635517 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.635536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.635570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.635590 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.738544 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.738611 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.738622 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.738640 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.738650 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.803742 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/0.log" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.803838 4960 generic.go:334] "Generic (PLEG): container finished" podID="c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e" containerID="a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012" exitCode=1 Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.803921 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerDied","Data":"a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.804748 4960 scope.go:117] "RemoveContainer" containerID="a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.819206 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.844158 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.844644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.844663 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.844697 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.844717 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.845863 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.862642 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.878246 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.903623 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.923360 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.939171 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.947909 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.947942 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.947954 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.947998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.948013 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.968207 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.981958 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4960]: I1002 07:17:29.995755 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.012297 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"2025-10-02T07:16:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667\\\\n2025-10-02T07:16:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667 to /host/opt/cni/bin/\\\\n2025-10-02T07:16:44Z [verbose] multus-daemon started\\\\n2025-10-02T07:16:44Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.025681 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.037588 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.050670 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.050737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.050757 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.050783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.050801 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.052911 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.073767 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.087639 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.097312 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.107016 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.153645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.153729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.153744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.153770 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.153788 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.257400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.257492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.257505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.257532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.257547 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.329928 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.330063 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:30 crc kubenswrapper[4960]: E1002 07:17:30.330202 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.330393 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:30 crc kubenswrapper[4960]: E1002 07:17:30.330422 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:30 crc kubenswrapper[4960]: E1002 07:17:30.330687 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.356066 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.360717 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.360782 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.360813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.360853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.360883 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.373246 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.392634 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.414515 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.427946 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.444017 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.456547 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.463791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.463847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.463864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.463886 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.463900 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.474127 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.493883 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.512874 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.531822 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.559258 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.567822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.567900 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.567925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.567958 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.568011 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.577212 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.591088 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.616493 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.632477 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.648316 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.668593 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"2025-10-02T07:16:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667\\\\n2025-10-02T07:16:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667 to /host/opt/cni/bin/\\\\n2025-10-02T07:16:44Z [verbose] multus-daemon started\\\\n2025-10-02T07:16:44Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.671315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.671350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.671360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.671376 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.671388 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.775188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.775257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.775273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.775296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.775310 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.810838 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/0.log" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.810959 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerStarted","Data":"f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.828104 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.845148 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.865284 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.881437 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.881485 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.881497 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.881515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.881525 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.884196 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.899206 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.919259 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.934510 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.958991 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.982764 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.984950 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.985006 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.985021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.985042 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.985056 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4960]: I1002 07:17:30.997696 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:30Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.018510 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"2025-10-02T07:16:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667\\\\n2025-10-02T07:16:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667 to /host/opt/cni/bin/\\\\n2025-10-02T07:16:44Z [verbose] multus-daemon started\\\\n2025-10-02T07:16:44Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.051505 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.066779 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.086608 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.088606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.088668 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.088678 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.088696 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.088708 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.110487 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.130721 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.152641 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.176514 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:31Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.192175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.192280 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.192305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.192341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.192366 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.295483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.295545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.295560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.295583 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.295596 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.329018 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:31 crc kubenswrapper[4960]: E1002 07:17:31.329182 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.398856 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.398943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.398961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.399016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.399044 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.502713 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.502753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.502762 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.502779 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.502789 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.605101 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.605221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.605236 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.605256 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.605266 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.707883 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.708021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.708039 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.708065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.708079 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.813330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.813410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.813429 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.813459 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.813481 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.915948 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.916007 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.916020 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.916040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4960]: I1002 07:17:31.916051 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.018209 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.018282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.018301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.018331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.018352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.121651 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.121711 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.121720 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.121739 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.121752 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.224464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.224520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.224534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.224560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.224577 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.327351 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.327419 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.327433 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.327459 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.327475 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.329564 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.329609 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:32 crc kubenswrapper[4960]: E1002 07:17:32.329764 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.329768 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:32 crc kubenswrapper[4960]: E1002 07:17:32.329905 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:32 crc kubenswrapper[4960]: E1002 07:17:32.330151 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.430271 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.430312 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.430322 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.430338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.430349 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.533817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.533870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.533882 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.533905 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.533920 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.637767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.637826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.637846 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.637871 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.637889 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.741490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.741539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.741555 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.741572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.741589 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.871644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.871679 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.871688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.871705 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.871716 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.976320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.976366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.976378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.976400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4960]: I1002 07:17:32.976411 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.080300 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.080364 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.080381 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.080405 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.080423 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.184936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.185008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.185022 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.185041 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.185052 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.289631 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.289700 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.289720 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.289756 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.289778 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.330078 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:33 crc kubenswrapper[4960]: E1002 07:17:33.330373 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.392750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.392835 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.392862 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.392891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.392910 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.497129 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.497199 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.497218 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.497248 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.497268 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.600627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.600716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.600735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.601232 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.601294 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.705845 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.705933 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.705957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.706029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.706057 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.809347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.809381 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.809389 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.809403 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.809412 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.912164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.912219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.912236 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.912258 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4960]: I1002 07:17:33.912273 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.015213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.015267 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.015279 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.015301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.015314 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.119241 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.119314 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.119326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.119356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.119372 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.223016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.223101 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.223117 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.223148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.223165 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.334529 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:34 crc kubenswrapper[4960]: E1002 07:17:34.334795 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.335254 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:34 crc kubenswrapper[4960]: E1002 07:17:34.335388 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.335647 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:34 crc kubenswrapper[4960]: E1002 07:17:34.335774 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.346855 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.346936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.346959 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.347027 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.347053 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.449732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.449776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.449795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.449811 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.449821 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.565342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.565477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.565502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.565602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.565661 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.668771 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.668814 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.668826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.668842 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.668889 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.772348 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.772413 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.772424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.772446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.772458 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.875207 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.875276 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.875294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.875324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.875343 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.979142 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.979191 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.979199 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.979216 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4960]: I1002 07:17:34.979227 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.083152 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.083208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.083226 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.083255 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.083274 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.187131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.187191 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.187200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.187219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.187230 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.290867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.290950 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.291018 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.291055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.291082 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.330024 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:35 crc kubenswrapper[4960]: E1002 07:17:35.330321 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.330352 4960 scope.go:117] "RemoveContainer" containerID="778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.394686 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.394729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.394741 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.394758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.394768 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.498114 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.498150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.498161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.498181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.498194 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.601024 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.601240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.601299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.601421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.601477 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.705168 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.705256 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.705282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.705311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.705335 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.785301 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.809436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.809512 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.809532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.809565 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.809586 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.839427 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/2.log" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.845321 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.846088 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.887495 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.906411 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.912638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.912688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.912703 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.912729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.912746 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.922851 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.937421 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.950182 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.964386 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.978127 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:35 crc kubenswrapper[4960]: I1002 07:17:35.996879 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:35Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.012492 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.015582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.015626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.015639 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.015660 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.015675 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.031637 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.046773 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.065751 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.080152 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.094247 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"2025-10-02T07:16:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667\\\\n2025-10-02T07:16:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667 to /host/opt/cni/bin/\\\\n2025-10-02T07:16:44Z [verbose] multus-daemon started\\\\n2025-10-02T07:16:44Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.124963 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.125050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.125070 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.125102 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.125122 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.138093 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.155660 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.166694 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.189907 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.228410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.228477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.228490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.228511 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.228526 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.329741 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.329743 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.329809 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:36 crc kubenswrapper[4960]: E1002 07:17:36.330012 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:36 crc kubenswrapper[4960]: E1002 07:17:36.330320 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:36 crc kubenswrapper[4960]: E1002 07:17:36.330446 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.332043 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.332072 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.332083 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.332104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.332117 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.435040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.435093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.435103 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.435120 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.435132 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.538879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.538925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.538936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.538953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.538968 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.641928 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.642032 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.642076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.642111 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.642132 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.746577 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.746704 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.746732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.746773 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.746803 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.850551 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.850603 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.850623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.850652 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.850672 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.854063 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/3.log" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.855277 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/2.log" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.860276 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" exitCode=1 Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.860325 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.860364 4960 scope.go:117] "RemoveContainer" containerID="778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.861679 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:17:36 crc kubenswrapper[4960]: E1002 07:17:36.861955 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.882428 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.902492 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.927727 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.947773 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.953114 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.953202 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.953230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.953271 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.953300 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.964177 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4960]: I1002 07:17:36.987217 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.010246 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.029405 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.052816 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"2025-10-02T07:16:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667\\\\n2025-10-02T07:16:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667 to /host/opt/cni/bin/\\\\n2025-10-02T07:16:44Z [verbose] multus-daemon started\\\\n2025-10-02T07:16:44Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.057402 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.057446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.057464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.057492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.057514 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.085624 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://778e664cdcca3a0dffc167060e27888c02efcfb1ce0c122a638bdc38f9f55f92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:06Z\\\",\\\"message\\\":\\\"ator/network-operator-58b4c7f79c-55gtf openshift-dns/node-resolver-bqvdt openshift-network-diagnostics/network-check-target-xd92c openshift-etcd/etcd-crc openshift-image-registry/node-ca-46hxp openshift-ovn-kubernetes/ovnkube-node-8dmqg openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-machine-config-operator/machine-config-daemon-rfcbg]\\\\nI1002 07:17:06.612192 6617 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1002 07:17:06.612194 6617 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:06Z is after 2025-08\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"ending *v1.Pod event handler 6 for removal\\\\nI1002 07:17:36.324711 6972 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:36.324706 6972 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:36.324739 6972 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:17:36.324747 6972 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:36.324722 6972 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:36.324754 6972 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:36.324803 6972 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:36.324819 6972 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:36.324832 6972 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:36.324841 6972 factory.go:656] Stopping watch factory\\\\nI1002 07:17:36.324866 6972 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:36.324966 6972 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 07:17:36.325099 6972 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 07:17:36.325142 6972 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:17:36.325175 6972 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:17:36.325247 6972 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.110063 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.128422 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.145777 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.162462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.162521 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.162534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.162564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.162582 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.166566 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.186933 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.201823 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.221556 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.239665 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.266906 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.266949 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.266962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.267012 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.267027 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.329234 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.329433 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.370722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.370785 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.370805 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.370830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.370850 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.427038 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.427073 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.427085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.427104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.427118 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.457197 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.464143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.464199 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.464220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.464250 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.464268 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.488423 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.495345 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.495412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.495432 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.495465 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.495485 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.518409 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.524786 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.524885 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.524918 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.524961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.525023 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.546629 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.552395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.552460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.552479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.552505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.552524 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.575414 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"903088be-089c-4b2a-9679-59229550832d\\\",\\\"systemUUID\\\":\\\"82b394c9-1025-4b62-b6ab-7caff0ff113f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.575681 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.578387 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.578462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.578481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.578510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.578529 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.682866 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.682953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.683027 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.683063 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.683082 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.787161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.787214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.787224 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.787242 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.787253 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.869234 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/3.log" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.875905 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:17:37 crc kubenswrapper[4960]: E1002 07:17:37.876226 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.890207 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.890286 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.890306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.890337 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.890364 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.900507 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.920550 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.944614 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.963330 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.994025 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.994082 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.994091 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.994115 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4960]: I1002 07:17:37.994131 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.004213 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.025307 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.042629 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.064435 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"2025-10-02T07:16:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667\\\\n2025-10-02T07:16:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667 to /host/opt/cni/bin/\\\\n2025-10-02T07:16:44Z [verbose] multus-daemon started\\\\n2025-10-02T07:16:44Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.090613 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"ending *v1.Pod event handler 6 for removal\\\\nI1002 07:17:36.324711 6972 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:36.324706 6972 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:36.324739 6972 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:17:36.324747 6972 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:36.324722 6972 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:36.324754 6972 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:36.324803 6972 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:36.324819 6972 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:36.324832 6972 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:36.324841 6972 factory.go:656] Stopping watch factory\\\\nI1002 07:17:36.324866 6972 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:36.324966 6972 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 07:17:36.325099 6972 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 07:17:36.325142 6972 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:17:36.325175 6972 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:17:36.325247 6972 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.098290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.098395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.098426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.098467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.098497 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.113059 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.128733 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.144211 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.163002 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.182306 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.198950 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.202211 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.202264 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.202281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.202305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.202319 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.217689 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.236250 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.254541 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.305391 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.305507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.305527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.305576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.305597 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.329517 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.329618 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.329558 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:38 crc kubenswrapper[4960]: E1002 07:17:38.329780 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:38 crc kubenswrapper[4960]: E1002 07:17:38.329889 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:38 crc kubenswrapper[4960]: E1002 07:17:38.330047 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.408844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.408905 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.408927 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.408961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.409019 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.512457 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.512524 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.512548 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.512578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.512605 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.616435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.616501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.616526 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.616558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.616581 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.719956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.720080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.720112 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.720148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.720177 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.824752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.825015 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.825055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.825131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.825165 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.929202 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.929272 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.929286 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.929306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4960]: I1002 07:17:38.929319 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.032435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.032477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.032488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.032504 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.032516 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.136076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.136139 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.136163 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.136189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.136205 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.239196 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.239249 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.239264 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.239286 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.239300 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.329780 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:39 crc kubenswrapper[4960]: E1002 07:17:39.330075 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.342755 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.342839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.342860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.342897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.342919 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.447098 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.447174 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.447197 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.447224 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.447242 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.550907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.551019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.551045 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.551079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.551105 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.655374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.655461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.655482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.655515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.655538 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.759860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.759927 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.759947 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.760009 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.760029 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.866019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.866085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.866103 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.866134 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.866153 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.969371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.969450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.969473 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.969504 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4960]: I1002 07:17:39.969525 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.073593 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.073666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.073694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.073726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.073745 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.176753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.176824 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.176843 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.176874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.176896 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.280757 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.280841 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.280865 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.280899 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.280925 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.329493 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.329536 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.329635 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:40 crc kubenswrapper[4960]: E1002 07:17:40.329753 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:40 crc kubenswrapper[4960]: E1002 07:17:40.329851 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:40 crc kubenswrapper[4960]: E1002 07:17:40.330248 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.346344 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.354680 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd1023a-fa1f-4d13-bde3-27af3d63bcad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0468f2df7a6f673d7f9281d3d06356835ce84ae6ed44097d2cb7386dfccb824f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe955c7d3ef8a805efa585f7ccc000838afb1567c51c09bb5f09ad72b6dad045\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9180ba879809af198065477ebcb6c2c8cfd8eedfdf0d54956bbe4937bd8efd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7973196a6d1152ea89210eaa8dd1e16b25bd1fe81fad76faa96e773197904d9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09c6b115c9d665cad097e3d4e636ff780ba5b42a80b707fb4879c9849c57eda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426a211d81de922903add10b544c43a3831ac3cfdaffed04aa8c99b283ccecff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.372450 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abf6f251-f711-44b9-9d4d-84eb2ccd1ded\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa07c34a1df611f1a6617cd44671cf3feaa5ab61bbb0120b9616d048f70800ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f74124f6cd927d623b4f5a125c25e3eed88ac2c9d410fea02c86016c0ff46960\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77990a8e503b705fa632212cad7fdde3cc4ddb2c11dcb8cb5c144ac4462bf6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c96daecf47000a9edc4376a09d2b7918f64ac207434bb7d8c35d26d882e53a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.385097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.385144 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.385156 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.385177 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.385190 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.389609 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc2a9d4090f7e8733adf2187b2e17ee7beedbefee4ee27f1f2d13fa22ea1980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42ce525a55c9a110d990949507bc764eee58f131a21beabac30dd7a7bcde9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.416695 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6bf9ea2-e784-43e4-a08a-7bcd491c805a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://164880be187b878c677716473bed8fc3a118b35a02a5438fdd27b4fbb005ed44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ce192d6ce3fac3ae4c55c900d51bd4d2877235d31103f16a8a7702d511f3daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9929w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rr46t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.443851 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1a53f0-2986-4090-a3c0-c7cac3b3fb57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ee67009a2931ed563c1d7afb238b48706d3641214aa1816bad7c8a8ade7525d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62bf49408d0c6a696831571514912f0e15d286cc6513671e622d62a1fd7f0908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0678bda53121742ea9514b55b19f9d7329058ba1e2f4a1d3e306a22c4af31806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9047bf64b0548cd6db42e92c60ad981a22d2d1c33e98dee7c5cdd65904871e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb08237bb95cbddbcd3725bb544e9db278801b2222c0ab556962193fbdd007c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a379fdc74d2bc57f8996c827af2a44962863edfedceb2d5a9e6391700aa0e529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a91ba7206bed212080768497c3c42b96e80e86f80b6e9867a92fae17bca3fcbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61a8cbeafec4d13154a35d8627e9b0107f73354b2d7d55bf161d57693136c407\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.462646 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.482150 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.488245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.488299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.488310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.488330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.488345 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.507615 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2dqxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:29Z\\\",\\\"message\\\":\\\"2025-10-02T07:16:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667\\\\n2025-10-02T07:16:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f5bff34-0f7b-4973-aebc-c1aaf47c9667 to /host/opt/cni/bin/\\\\n2025-10-02T07:16:44Z [verbose] multus-daemon started\\\\n2025-10-02T07:16:44Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdwn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2dqxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.534482 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ab5426-0159-4610-b378-f4c2e20b82b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"ending *v1.Pod event handler 6 for removal\\\\nI1002 07:17:36.324711 6972 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:36.324706 6972 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:36.324739 6972 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 07:17:36.324747 6972 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:36.324722 6972 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:36.324754 6972 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:36.324803 6972 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:36.324819 6972 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:36.324832 6972 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:36.324841 6972 factory.go:656] Stopping watch factory\\\\nI1002 07:17:36.324866 6972 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:36.324966 6972 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 07:17:36.325099 6972 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 07:17:36.325142 6972 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:17:36.325175 6972 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 07:17:36.325247 6972 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dxd2c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8dmqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.562485 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vm98j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb4216ca-0206-4794-bb05-efabd0dc96cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90b461a3bb402a7c37f8c1d74d548f8c19c8e27f423ae951695e05979457c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f667c7fc1ffc704590d63212f7d41f2a6efb2a5c087105b8d28ad709dbfcb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fe43664f39331c15383f54a40cb18a23daf31463d2f7137d16a531f5a57f3f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4069c66f6acf1a9794a9a9ee1f6f93451dc9f5d20270e9e471d2c51209cccfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fffaae9a4dd5c87c06136a9cf938f647bf29db981e68ce533c7cf0104d8944a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a5d62f3e7687e01e6175123e410c7a1a19d2485cd05d9ac3f6bc4db53c677f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80258bcf101127405b9d44a0be12c89db98d844f0319777888322806bbac2cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n84cc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vm98j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.579252 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctnf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hhvkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.592602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.592708 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.592736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.592777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.592803 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.597160 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1ed00c4-01cd-42ca-b97f-1132e08a58b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2393251c8c18fbb0c1ee10669e785398ab1d41f93b400f0474370a4897d7e65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f52b4b0f39352d3328149a7dc9fa2feba6d4f8ef54b4d6c23d051d7afd5ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ea77767d554746f1d8294c4d54bdbe39a96a159ad792d5835f13b7009a84fd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38fe1768759690afaca2cff8900e059976548d294de035e5f045f56a07856f33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.619312 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da4651e406d196da87393c3d1aea404386321fe29d7ecb73074522d4f4994bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.639205 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.654731 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bqvdt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e523ebca-7bf0-4904-9671-8c0381b01be8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://411d6edb255461d484839593ce8abe2a0b5eb2bcfe75f7e124cefc503d05d586\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bqvdt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.669046 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a813e5e1-abad-433b-a1b3-24cab8119638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1916228ceee9d0dddbeaa341e53f542e0ccf0dbc3309540ba453caf93cf7000\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dqjb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rfcbg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.688060 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-46hxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8573c7cf-f32d-4f64-a857-5136816b5173\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c45d5919b3645da63252457f44a5e1668bdb997963c09e26d45c868b2636698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8qkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-46hxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.696192 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.696258 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.696278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.696310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.696332 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.709613 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc8bf4f36215d19fef08b736819c2ae85cecdbcfb2aa13f7740554b4a69b3ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:40Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.799212 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.799267 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.799281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.799301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.799314 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.902213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.902872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.902923 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.902955 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4960]: I1002 07:17:40.903011 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.006625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.006691 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.006712 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.006744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.006765 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.109886 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.110519 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.110602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.110692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.110783 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.213315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.213426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.213454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.213488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.213510 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.316813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.316866 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.316878 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.316897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.316909 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.330227 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:41 crc kubenswrapper[4960]: E1002 07:17:41.330340 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.418767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.419059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.419274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.419366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.419452 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.522137 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.522181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.522190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.522206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.522220 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.624353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.624403 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.624415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.624442 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.624453 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.727392 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.727440 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.727452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.727468 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.727484 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.837507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.837595 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.837626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.837673 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.837699 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.941799 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.941843 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.941853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.941870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4960]: I1002 07:17:41.941880 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.044994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.045035 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.045044 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.045061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.045076 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.146691 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.146727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.146737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.146753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.146764 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.249415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.249715 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.249817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.249916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.250032 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.329295 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.329311 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.329335 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:42 crc kubenswrapper[4960]: E1002 07:17:42.329596 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:42 crc kubenswrapper[4960]: E1002 07:17:42.329548 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:42 crc kubenswrapper[4960]: E1002 07:17:42.329693 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.352288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.352518 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.352593 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.352671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.352740 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.454746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.454794 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.454804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.454822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.454833 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.557473 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.557544 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.557557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.557573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.557582 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.661725 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.661801 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.661823 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.661857 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.661883 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.765869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.765960 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.766021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.766336 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.766406 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.869522 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.869570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.869581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.869600 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.869611 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.974066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.974362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.974638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.974740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4960]: I1002 07:17:42.974803 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.078122 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.078162 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.078171 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.078186 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.078199 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.181111 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.181195 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.181223 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.181259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.181284 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.284209 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.284246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.284261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.284279 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.284290 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.329652 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.329945 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.386515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.386559 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.386570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.386586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.386597 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.489850 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.489920 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.489940 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.489971 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.490020 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.505458 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.505607 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.505642 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.505671 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.505781 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.505745938 +0000 UTC m=+148.537692265 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.505816 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.505855 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.506056 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.506021835 +0000 UTC m=+148.537968122 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.506126 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.505880 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.506290 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.506265982 +0000 UTC m=+148.538212309 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.506204 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.506439 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.506416025 +0000 UTC m=+148.538362352 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.593596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.593645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.593655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.593675 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.593687 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.607286 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.607434 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.607459 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.607470 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:43 crc kubenswrapper[4960]: E1002 07:17:43.607521 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.60750044 +0000 UTC m=+148.639446727 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.696745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.696786 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.696795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.696810 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.696821 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.799681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.799717 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.799728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.799745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.799754 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.902181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.902222 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.902231 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.902246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4960]: I1002 07:17:43.902255 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.006030 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.006080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.006097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.006116 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.006129 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.109480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.109527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.109537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.109557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.109566 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.212082 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.212142 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.212163 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.212190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.212215 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.318440 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.318503 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.318517 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.318541 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.318559 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.329538 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.329568 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.329734 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:44 crc kubenswrapper[4960]: E1002 07:17:44.329952 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:44 crc kubenswrapper[4960]: E1002 07:17:44.330819 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:44 crc kubenswrapper[4960]: E1002 07:17:44.330926 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.422044 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.422450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.422602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.422747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.422873 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.527373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.527464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.527489 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.527525 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.527553 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.630354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.630429 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.630443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.630464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.630474 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.733398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.733449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.733460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.733479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.733492 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.836615 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.836703 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.836721 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.836748 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.836768 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.940804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.940849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.940859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.940879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4960]: I1002 07:17:44.940889 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.043934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.044297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.044362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.044433 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.044496 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.147467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.147546 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.147566 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.147597 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.147618 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.250781 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.250842 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.250864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.250890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.250909 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.329664 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:45 crc kubenswrapper[4960]: E1002 07:17:45.329824 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.353632 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.353678 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.353692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.353710 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.353728 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.456889 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.456959 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.457005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.457035 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.457056 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.560164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.560305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.560331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.560368 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.560387 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.663653 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.663733 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.663757 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.663798 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.663823 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.767493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.767565 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.767581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.767602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.767620 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.871110 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.871166 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.871180 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.871200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.871214 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.975492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.975737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.975771 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.975809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4960]: I1002 07:17:45.975840 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.079935 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.080593 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.080608 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.080636 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.080660 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.184390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.184440 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.184451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.184472 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.184487 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.287080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.287147 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.287165 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.287194 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.287214 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.328904 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.329081 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:46 crc kubenswrapper[4960]: E1002 07:17:46.329185 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.329319 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:46 crc kubenswrapper[4960]: E1002 07:17:46.329641 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:46 crc kubenswrapper[4960]: E1002 07:17:46.330105 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.390428 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.390506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.390529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.390561 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.390584 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.494498 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.494586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.494606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.494639 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.494668 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.597591 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.597647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.597659 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.597678 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.597702 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.705738 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.706222 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.706441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.706580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.706718 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.810306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.810373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.810392 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.810449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.810468 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.920072 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.920168 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.920193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.920230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4960]: I1002 07:17:46.920256 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.024405 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.024471 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.024484 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.024508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.024533 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.128494 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.128557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.128575 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.128601 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.128625 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.231932 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.232080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.232106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.232137 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.232155 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.329636 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:47 crc kubenswrapper[4960]: E1002 07:17:47.329826 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.335057 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.335170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.335206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.335239 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.335261 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.439388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.439441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.439451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.439470 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.439486 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.542708 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.542771 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.542783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.542807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.542824 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.646258 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.646318 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.646333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.646356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.646372 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.749077 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.749155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.749172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.749202 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.749221 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.768523 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.768565 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.768582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.768601 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.768618 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.841003 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j"] Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.841713 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.844172 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.844239 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.844659 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.847146 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.929107 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rr46t" podStartSLOduration=65.92906319 podStartE2EDuration="1m5.92906319s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:47.928702701 +0000 UTC m=+88.960648988" watchObservedRunningTime="2025-10-02 07:17:47.92906319 +0000 UTC m=+88.961009527" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.958370 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.958440 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.958480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.958512 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.958568 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:47 crc kubenswrapper[4960]: I1002 07:17:47.975356 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=7.975324946 podStartE2EDuration="7.975324946s" podCreationTimestamp="2025-10-02 07:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:47.947770738 +0000 UTC m=+88.979717045" watchObservedRunningTime="2025-10-02 07:17:47.975324946 +0000 UTC m=+89.007271253" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.005246 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.005219622 podStartE2EDuration="1m9.005219622s" podCreationTimestamp="2025-10-02 07:16:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:47.975857259 +0000 UTC m=+89.007803546" watchObservedRunningTime="2025-10-02 07:17:48.005219622 +0000 UTC m=+89.037165909" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.029614 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=61.029585021 podStartE2EDuration="1m1.029585021s" podCreationTimestamp="2025-10-02 07:16:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.006061253 +0000 UTC m=+89.038007610" watchObservedRunningTime="2025-10-02 07:17:48.029585021 +0000 UTC m=+89.061531308" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.029911 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2dqxx" podStartSLOduration=67.029906018 podStartE2EDuration="1m7.029906018s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.029831207 +0000 UTC m=+89.061777534" watchObservedRunningTime="2025-10-02 07:17:48.029906018 +0000 UTC m=+89.061852295" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.060060 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.060601 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.060778 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.060932 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.061146 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.061328 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.061406 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.062897 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.072062 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.096061 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39f5e5fc-608c-4c84-8a1e-c57bc9abea0e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5jw5j\" (UID: \"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.107827 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vm98j" podStartSLOduration=67.107793694 podStartE2EDuration="1m7.107793694s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.107265141 +0000 UTC m=+89.139211468" watchObservedRunningTime="2025-10-02 07:17:48.107793694 +0000 UTC m=+89.139740001" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.154790 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=67.154758537 podStartE2EDuration="1m7.154758537s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.154569723 +0000 UTC m=+89.186516060" watchObservedRunningTime="2025-10-02 07:17:48.154758537 +0000 UTC m=+89.186704824" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.161691 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" Oct 02 07:17:48 crc kubenswrapper[4960]: W1002 07:17:48.185845 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39f5e5fc_608c_4c84_8a1e_c57bc9abea0e.slice/crio-2e96ee6481644dcec88a188fd60f28830b09213189dc6209e91eb0f530e7642a WatchSource:0}: Error finding container 2e96ee6481644dcec88a188fd60f28830b09213189dc6209e91eb0f530e7642a: Status 404 returned error can't find the container with id 2e96ee6481644dcec88a188fd60f28830b09213189dc6209e91eb0f530e7642a Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.220025 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-bqvdt" podStartSLOduration=67.219951965 podStartE2EDuration="1m7.219951965s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.217762981 +0000 UTC m=+89.249709268" watchObservedRunningTime="2025-10-02 07:17:48.219951965 +0000 UTC m=+89.251898292" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.255204 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podStartSLOduration=67.255171835 podStartE2EDuration="1m7.255171835s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.242240332 +0000 UTC m=+89.274186619" watchObservedRunningTime="2025-10-02 07:17:48.255171835 +0000 UTC m=+89.287118162" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.256123 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-46hxp" podStartSLOduration=67.256115258 podStartE2EDuration="1m7.256115258s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.255399671 +0000 UTC m=+89.287345958" watchObservedRunningTime="2025-10-02 07:17:48.256115258 +0000 UTC m=+89.288061585" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.287934 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=32.287914433 podStartE2EDuration="32.287914433s" podCreationTimestamp="2025-10-02 07:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.272509388 +0000 UTC m=+89.304455675" watchObservedRunningTime="2025-10-02 07:17:48.287914433 +0000 UTC m=+89.319860720" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.329418 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.330372 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.331118 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:48 crc kubenswrapper[4960]: E1002 07:17:48.330598 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:48 crc kubenswrapper[4960]: E1002 07:17:48.331299 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:48 crc kubenswrapper[4960]: E1002 07:17:48.331092 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.913683 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" event={"ID":"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e","Type":"ContainerStarted","Data":"8100090ba1c0dfad1cf6f059d502eb8e1c3dc91c25cbd30119069920b4ecc5f7"} Oct 02 07:17:48 crc kubenswrapper[4960]: I1002 07:17:48.913758 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" event={"ID":"39f5e5fc-608c-4c84-8a1e-c57bc9abea0e","Type":"ContainerStarted","Data":"2e96ee6481644dcec88a188fd60f28830b09213189dc6209e91eb0f530e7642a"} Oct 02 07:17:49 crc kubenswrapper[4960]: I1002 07:17:49.329156 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:49 crc kubenswrapper[4960]: E1002 07:17:49.330070 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:49 crc kubenswrapper[4960]: I1002 07:17:49.330745 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:17:49 crc kubenswrapper[4960]: E1002 07:17:49.331234 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" Oct 02 07:17:50 crc kubenswrapper[4960]: I1002 07:17:50.329358 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:50 crc kubenswrapper[4960]: I1002 07:17:50.329471 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:50 crc kubenswrapper[4960]: I1002 07:17:50.333048 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:50 crc kubenswrapper[4960]: E1002 07:17:50.333196 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:50 crc kubenswrapper[4960]: E1002 07:17:50.333579 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:50 crc kubenswrapper[4960]: E1002 07:17:50.333056 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:51 crc kubenswrapper[4960]: I1002 07:17:51.329738 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:51 crc kubenswrapper[4960]: E1002 07:17:51.330040 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:52 crc kubenswrapper[4960]: I1002 07:17:52.329141 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:52 crc kubenswrapper[4960]: I1002 07:17:52.329339 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:52 crc kubenswrapper[4960]: I1002 07:17:52.329544 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:52 crc kubenswrapper[4960]: E1002 07:17:52.329713 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:52 crc kubenswrapper[4960]: E1002 07:17:52.330100 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:52 crc kubenswrapper[4960]: E1002 07:17:52.330335 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:53 crc kubenswrapper[4960]: I1002 07:17:53.329361 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:53 crc kubenswrapper[4960]: E1002 07:17:53.330186 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:54 crc kubenswrapper[4960]: I1002 07:17:54.329244 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:54 crc kubenswrapper[4960]: I1002 07:17:54.329282 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:54 crc kubenswrapper[4960]: E1002 07:17:54.329420 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:54 crc kubenswrapper[4960]: I1002 07:17:54.329454 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:54 crc kubenswrapper[4960]: E1002 07:17:54.329609 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:54 crc kubenswrapper[4960]: E1002 07:17:54.329716 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:55 crc kubenswrapper[4960]: I1002 07:17:55.329129 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:55 crc kubenswrapper[4960]: E1002 07:17:55.329253 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:56 crc kubenswrapper[4960]: I1002 07:17:56.330491 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:56 crc kubenswrapper[4960]: I1002 07:17:56.330598 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:56 crc kubenswrapper[4960]: E1002 07:17:56.331463 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:56 crc kubenswrapper[4960]: E1002 07:17:56.331397 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:56 crc kubenswrapper[4960]: I1002 07:17:56.330813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:56 crc kubenswrapper[4960]: E1002 07:17:56.331534 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:57 crc kubenswrapper[4960]: I1002 07:17:57.329964 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:57 crc kubenswrapper[4960]: E1002 07:17:57.330162 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:58 crc kubenswrapper[4960]: I1002 07:17:58.329194 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:58 crc kubenswrapper[4960]: E1002 07:17:58.329326 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:58 crc kubenswrapper[4960]: I1002 07:17:58.329511 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:17:58 crc kubenswrapper[4960]: E1002 07:17:58.329568 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:17:58 crc kubenswrapper[4960]: I1002 07:17:58.330016 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:58 crc kubenswrapper[4960]: E1002 07:17:58.330262 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:59 crc kubenswrapper[4960]: I1002 07:17:59.329681 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:59 crc kubenswrapper[4960]: E1002 07:17:59.330101 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:00 crc kubenswrapper[4960]: I1002 07:18:00.211764 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:00 crc kubenswrapper[4960]: E1002 07:18:00.212153 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:18:00 crc kubenswrapper[4960]: E1002 07:18:00.212296 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs podName:09886fa3-04f6-4e09-acc1-b983cbdfd6ba nodeName:}" failed. No retries permitted until 2025-10-02 07:19:04.212255213 +0000 UTC m=+165.244201540 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs") pod "network-metrics-daemon-hhvkr" (UID: "09886fa3-04f6-4e09-acc1-b983cbdfd6ba") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:18:00 crc kubenswrapper[4960]: I1002 07:18:00.329121 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:00 crc kubenswrapper[4960]: I1002 07:18:00.329264 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:00 crc kubenswrapper[4960]: E1002 07:18:00.329316 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:00 crc kubenswrapper[4960]: E1002 07:18:00.329489 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:00 crc kubenswrapper[4960]: I1002 07:18:00.330042 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:00 crc kubenswrapper[4960]: E1002 07:18:00.331543 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:01 crc kubenswrapper[4960]: I1002 07:18:01.329797 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:01 crc kubenswrapper[4960]: E1002 07:18:01.330057 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:02 crc kubenswrapper[4960]: I1002 07:18:02.329856 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:02 crc kubenswrapper[4960]: I1002 07:18:02.329911 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:02 crc kubenswrapper[4960]: E1002 07:18:02.329999 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:02 crc kubenswrapper[4960]: E1002 07:18:02.330075 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:02 crc kubenswrapper[4960]: I1002 07:18:02.330136 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:02 crc kubenswrapper[4960]: E1002 07:18:02.330323 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:03 crc kubenswrapper[4960]: I1002 07:18:03.329048 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:03 crc kubenswrapper[4960]: E1002 07:18:03.329637 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:03 crc kubenswrapper[4960]: I1002 07:18:03.329915 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:18:03 crc kubenswrapper[4960]: E1002 07:18:03.330164 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8dmqg_openshift-ovn-kubernetes(41ab5426-0159-4610-b378-f4c2e20b82b9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" Oct 02 07:18:04 crc kubenswrapper[4960]: I1002 07:18:04.329194 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:04 crc kubenswrapper[4960]: I1002 07:18:04.329194 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:04 crc kubenswrapper[4960]: E1002 07:18:04.329310 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:04 crc kubenswrapper[4960]: I1002 07:18:04.329365 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:04 crc kubenswrapper[4960]: E1002 07:18:04.329830 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:04 crc kubenswrapper[4960]: E1002 07:18:04.329956 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:05 crc kubenswrapper[4960]: I1002 07:18:05.329731 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:05 crc kubenswrapper[4960]: E1002 07:18:05.330314 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:06 crc kubenswrapper[4960]: I1002 07:18:06.329239 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:06 crc kubenswrapper[4960]: I1002 07:18:06.329240 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:06 crc kubenswrapper[4960]: I1002 07:18:06.329285 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:06 crc kubenswrapper[4960]: E1002 07:18:06.329491 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:06 crc kubenswrapper[4960]: E1002 07:18:06.329610 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:06 crc kubenswrapper[4960]: E1002 07:18:06.329842 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:07 crc kubenswrapper[4960]: I1002 07:18:07.329100 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:07 crc kubenswrapper[4960]: E1002 07:18:07.329283 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:08 crc kubenswrapper[4960]: I1002 07:18:08.330142 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:08 crc kubenswrapper[4960]: I1002 07:18:08.330197 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:08 crc kubenswrapper[4960]: E1002 07:18:08.330300 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:08 crc kubenswrapper[4960]: E1002 07:18:08.330344 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:08 crc kubenswrapper[4960]: I1002 07:18:08.331194 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:08 crc kubenswrapper[4960]: E1002 07:18:08.331419 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:09 crc kubenswrapper[4960]: I1002 07:18:09.329500 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:09 crc kubenswrapper[4960]: E1002 07:18:09.329704 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:10 crc kubenswrapper[4960]: I1002 07:18:10.329124 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:10 crc kubenswrapper[4960]: I1002 07:18:10.329285 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:10 crc kubenswrapper[4960]: I1002 07:18:10.329312 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:10 crc kubenswrapper[4960]: E1002 07:18:10.330484 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:10 crc kubenswrapper[4960]: E1002 07:18:10.330648 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:10 crc kubenswrapper[4960]: E1002 07:18:10.330761 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:11 crc kubenswrapper[4960]: I1002 07:18:11.329409 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:11 crc kubenswrapper[4960]: E1002 07:18:11.329555 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:12 crc kubenswrapper[4960]: I1002 07:18:12.330177 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:12 crc kubenswrapper[4960]: I1002 07:18:12.330248 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:12 crc kubenswrapper[4960]: E1002 07:18:12.330349 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:12 crc kubenswrapper[4960]: I1002 07:18:12.330368 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:12 crc kubenswrapper[4960]: E1002 07:18:12.330422 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:12 crc kubenswrapper[4960]: E1002 07:18:12.330633 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:13 crc kubenswrapper[4960]: I1002 07:18:13.329044 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:13 crc kubenswrapper[4960]: E1002 07:18:13.329195 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:14 crc kubenswrapper[4960]: I1002 07:18:14.328836 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:14 crc kubenswrapper[4960]: I1002 07:18:14.328947 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:14 crc kubenswrapper[4960]: E1002 07:18:14.329030 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:14 crc kubenswrapper[4960]: I1002 07:18:14.328948 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:14 crc kubenswrapper[4960]: E1002 07:18:14.329260 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:14 crc kubenswrapper[4960]: E1002 07:18:14.329501 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:15 crc kubenswrapper[4960]: I1002 07:18:15.328839 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:15 crc kubenswrapper[4960]: E1002 07:18:15.329015 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.007075 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/1.log" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.007612 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/0.log" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.007685 4960 generic.go:334] "Generic (PLEG): container finished" podID="c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e" containerID="f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3" exitCode=1 Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.007725 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerDied","Data":"f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3"} Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.007766 4960 scope.go:117] "RemoveContainer" containerID="a46f3c4e12b0a9f85f372a65f4908f2e07c257a5cc20ea1224e6d15a0a1e5012" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.008692 4960 scope.go:117] "RemoveContainer" containerID="f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3" Oct 02 07:18:16 crc kubenswrapper[4960]: E1002 07:18:16.009229 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2dqxx_openshift-multus(c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e)\"" pod="openshift-multus/multus-2dqxx" podUID="c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.039407 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5jw5j" podStartSLOduration=95.039376215 podStartE2EDuration="1m35.039376215s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:17:48.93162931 +0000 UTC m=+89.963575597" watchObservedRunningTime="2025-10-02 07:18:16.039376215 +0000 UTC m=+117.071322532" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.329282 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.329406 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:16 crc kubenswrapper[4960]: I1002 07:18:16.329410 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:16 crc kubenswrapper[4960]: E1002 07:18:16.329541 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:16 crc kubenswrapper[4960]: E1002 07:18:16.329781 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:16 crc kubenswrapper[4960]: E1002 07:18:16.330068 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:17 crc kubenswrapper[4960]: I1002 07:18:17.014451 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/1.log" Oct 02 07:18:17 crc kubenswrapper[4960]: I1002 07:18:17.329086 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:17 crc kubenswrapper[4960]: E1002 07:18:17.329253 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:18 crc kubenswrapper[4960]: I1002 07:18:18.329157 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:18 crc kubenswrapper[4960]: I1002 07:18:18.329198 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:18 crc kubenswrapper[4960]: I1002 07:18:18.329479 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:18 crc kubenswrapper[4960]: E1002 07:18:18.329766 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:18 crc kubenswrapper[4960]: I1002 07:18:18.329946 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:18:18 crc kubenswrapper[4960]: E1002 07:18:18.329945 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:18 crc kubenswrapper[4960]: E1002 07:18:18.330178 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:19 crc kubenswrapper[4960]: I1002 07:18:19.023792 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/3.log" Oct 02 07:18:19 crc kubenswrapper[4960]: I1002 07:18:19.028244 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerStarted","Data":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} Oct 02 07:18:19 crc kubenswrapper[4960]: I1002 07:18:19.028775 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:18:19 crc kubenswrapper[4960]: I1002 07:18:19.057464 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podStartSLOduration=98.057439722 podStartE2EDuration="1m38.057439722s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:19.05659701 +0000 UTC m=+120.088543307" watchObservedRunningTime="2025-10-02 07:18:19.057439722 +0000 UTC m=+120.089386009" Oct 02 07:18:19 crc kubenswrapper[4960]: I1002 07:18:19.280692 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hhvkr"] Oct 02 07:18:19 crc kubenswrapper[4960]: I1002 07:18:19.280902 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:19 crc kubenswrapper[4960]: E1002 07:18:19.281141 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:19 crc kubenswrapper[4960]: I1002 07:18:19.366234 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:19 crc kubenswrapper[4960]: E1002 07:18:19.366410 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:20 crc kubenswrapper[4960]: I1002 07:18:20.329392 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:20 crc kubenswrapper[4960]: E1002 07:18:20.331879 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:20 crc kubenswrapper[4960]: I1002 07:18:20.333113 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:20 crc kubenswrapper[4960]: E1002 07:18:20.334243 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:20 crc kubenswrapper[4960]: E1002 07:18:20.347030 4960 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 07:18:20 crc kubenswrapper[4960]: E1002 07:18:20.499717 4960 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:18:21 crc kubenswrapper[4960]: I1002 07:18:21.329150 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:21 crc kubenswrapper[4960]: I1002 07:18:21.329217 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:21 crc kubenswrapper[4960]: E1002 07:18:21.329277 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:21 crc kubenswrapper[4960]: E1002 07:18:21.329388 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:22 crc kubenswrapper[4960]: I1002 07:18:22.329332 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:22 crc kubenswrapper[4960]: E1002 07:18:22.329478 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:22 crc kubenswrapper[4960]: I1002 07:18:22.329330 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:22 crc kubenswrapper[4960]: E1002 07:18:22.329719 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:23 crc kubenswrapper[4960]: I1002 07:18:23.329301 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:23 crc kubenswrapper[4960]: I1002 07:18:23.329368 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:23 crc kubenswrapper[4960]: E1002 07:18:23.329464 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:23 crc kubenswrapper[4960]: E1002 07:18:23.329568 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:24 crc kubenswrapper[4960]: I1002 07:18:24.329128 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:24 crc kubenswrapper[4960]: I1002 07:18:24.329167 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:24 crc kubenswrapper[4960]: E1002 07:18:24.329269 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:24 crc kubenswrapper[4960]: E1002 07:18:24.329406 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:25 crc kubenswrapper[4960]: I1002 07:18:25.329425 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:25 crc kubenswrapper[4960]: I1002 07:18:25.329425 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:25 crc kubenswrapper[4960]: E1002 07:18:25.329662 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:25 crc kubenswrapper[4960]: E1002 07:18:25.329544 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:25 crc kubenswrapper[4960]: E1002 07:18:25.501266 4960 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:18:26 crc kubenswrapper[4960]: I1002 07:18:26.329934 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:26 crc kubenswrapper[4960]: E1002 07:18:26.330158 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:26 crc kubenswrapper[4960]: I1002 07:18:26.330332 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:26 crc kubenswrapper[4960]: E1002 07:18:26.330487 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:27 crc kubenswrapper[4960]: I1002 07:18:27.329077 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:27 crc kubenswrapper[4960]: I1002 07:18:27.329124 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:27 crc kubenswrapper[4960]: E1002 07:18:27.329220 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:27 crc kubenswrapper[4960]: E1002 07:18:27.329343 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:28 crc kubenswrapper[4960]: I1002 07:18:28.329660 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:28 crc kubenswrapper[4960]: I1002 07:18:28.329660 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:28 crc kubenswrapper[4960]: E1002 07:18:28.329859 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:28 crc kubenswrapper[4960]: E1002 07:18:28.330086 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:29 crc kubenswrapper[4960]: I1002 07:18:29.329639 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:29 crc kubenswrapper[4960]: I1002 07:18:29.329781 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:29 crc kubenswrapper[4960]: E1002 07:18:29.329787 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:29 crc kubenswrapper[4960]: E1002 07:18:29.330070 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:29 crc kubenswrapper[4960]: I1002 07:18:29.330320 4960 scope.go:117] "RemoveContainer" containerID="f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3" Oct 02 07:18:30 crc kubenswrapper[4960]: I1002 07:18:30.072187 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/1.log" Oct 02 07:18:30 crc kubenswrapper[4960]: I1002 07:18:30.072927 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerStarted","Data":"58bee76173a69e8b6bf08b9a1abd3626ec971618f989710817a49de54f5243e2"} Oct 02 07:18:30 crc kubenswrapper[4960]: I1002 07:18:30.329347 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:30 crc kubenswrapper[4960]: I1002 07:18:30.329347 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:30 crc kubenswrapper[4960]: E1002 07:18:30.335270 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:30 crc kubenswrapper[4960]: E1002 07:18:30.335338 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:30 crc kubenswrapper[4960]: E1002 07:18:30.502079 4960 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:18:31 crc kubenswrapper[4960]: I1002 07:18:31.329471 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:31 crc kubenswrapper[4960]: I1002 07:18:31.329530 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:31 crc kubenswrapper[4960]: E1002 07:18:31.329615 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:31 crc kubenswrapper[4960]: E1002 07:18:31.329693 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:32 crc kubenswrapper[4960]: I1002 07:18:32.329449 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:32 crc kubenswrapper[4960]: I1002 07:18:32.329602 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:32 crc kubenswrapper[4960]: E1002 07:18:32.329641 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:32 crc kubenswrapper[4960]: E1002 07:18:32.329874 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:33 crc kubenswrapper[4960]: I1002 07:18:33.328871 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:33 crc kubenswrapper[4960]: E1002 07:18:33.329131 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:33 crc kubenswrapper[4960]: I1002 07:18:33.328871 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:33 crc kubenswrapper[4960]: E1002 07:18:33.329422 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:34 crc kubenswrapper[4960]: I1002 07:18:34.329226 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:34 crc kubenswrapper[4960]: I1002 07:18:34.329271 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:34 crc kubenswrapper[4960]: E1002 07:18:34.329519 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:34 crc kubenswrapper[4960]: E1002 07:18:34.329610 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:35 crc kubenswrapper[4960]: I1002 07:18:35.329418 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:35 crc kubenswrapper[4960]: I1002 07:18:35.329535 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:35 crc kubenswrapper[4960]: E1002 07:18:35.329557 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:35 crc kubenswrapper[4960]: E1002 07:18:35.329804 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hhvkr" podUID="09886fa3-04f6-4e09-acc1-b983cbdfd6ba" Oct 02 07:18:35 crc kubenswrapper[4960]: I1002 07:18:35.801159 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:18:36 crc kubenswrapper[4960]: I1002 07:18:36.329724 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:36 crc kubenswrapper[4960]: I1002 07:18:36.330593 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:36 crc kubenswrapper[4960]: I1002 07:18:36.332095 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 07:18:36 crc kubenswrapper[4960]: I1002 07:18:36.332142 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 07:18:36 crc kubenswrapper[4960]: I1002 07:18:36.333597 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 07:18:36 crc kubenswrapper[4960]: I1002 07:18:36.334085 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 07:18:37 crc kubenswrapper[4960]: I1002 07:18:37.329636 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:37 crc kubenswrapper[4960]: I1002 07:18:37.329718 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:18:37 crc kubenswrapper[4960]: I1002 07:18:37.331846 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 07:18:37 crc kubenswrapper[4960]: I1002 07:18:37.332450 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.816421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.872343 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.873445 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.879055 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.879497 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.879706 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.880134 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.880795 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.882171 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.882673 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.883561 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.883694 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.884900 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h8cqc"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.885336 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.886562 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lddcs"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.887445 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zcpbw"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.887819 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.887925 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.894678 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-89pc5"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.895784 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.896125 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.896679 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-rc25f"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.897493 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rc25f" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.899167 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.899285 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.899346 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.899555 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.899805 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.899938 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.900677 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.900887 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.901271 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.902045 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.902296 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.906254 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.911884 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vz5gr"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.915952 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916227 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fr4q7"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916738 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916277 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916321 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916550 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916577 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916618 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916813 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.916871 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.917288 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.921044 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jzg6w"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.921298 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.934954 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.936847 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.938846 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.939463 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.941569 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.954533 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.955373 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.957569 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8md6w"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.958043 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xxnvh"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.958466 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-smfhg"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.958797 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.963110 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.963358 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.965920 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966027 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966125 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966201 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966366 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966454 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966544 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966703 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966787 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966863 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.966956 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967069 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967078 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967207 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967243 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967289 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967383 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967405 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967480 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967568 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967576 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967714 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967725 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967869 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967916 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.968051 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.968201 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.968282 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969338 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969514 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969596 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969686 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969783 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969836 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969902 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.969949 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.970018 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.977840 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.977929 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.978296 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.978689 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.979390 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.979807 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.989788 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-service-ca-bundle\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.989848 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-trusted-ca-bundle\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.989876 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/604db804-d20c-4233-99d3-98690412d5cd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7r26m\" (UID: \"604db804-d20c-4233-99d3-98690412d5cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.989901 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.989926 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.989969 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.990031 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.990055 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg6v7\" (UniqueName: \"kubernetes.io/projected/ec915743-c847-49f8-9d0e-c800b9fb2f2a-kube-api-access-pg6v7\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.990211 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5dqd\" (UniqueName: \"kubernetes.io/projected/be880348-4fad-4b89-8e00-6dd336636f76-kube-api-access-l5dqd\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.967871 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.990311 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-config\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.990341 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.982123 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9lf9c"] Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.980503 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.980846 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.980896 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.981889 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.982127 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985262 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985281 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985337 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985380 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985417 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985632 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985674 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985801 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985837 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.991042 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985898 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985927 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.985949 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.986054 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.986261 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.986336 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.986380 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.986526 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.986535 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.987073 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.987154 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.987959 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.988231 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.988299 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.988523 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.994925 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.996958 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997055 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997092 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997125 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-client-ca\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997144 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2333c3b2-23cb-4e77-a88c-67f9118eb8de-trusted-ca\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997169 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfvlq\" (UniqueName: \"kubernetes.io/projected/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-kube-api-access-nfvlq\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997224 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7mfc\" (UniqueName: \"kubernetes.io/projected/7a93827e-e818-45a4-9276-bada4a616712-kube-api-access-z7mfc\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997252 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-encryption-config\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997286 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-node-pullsecrets\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997309 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-config\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997334 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-service-ca\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997358 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997380 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-serving-cert\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997406 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-config\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997426 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-config\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997445 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-config\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997466 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b130afa1-783b-4b51-98bb-567c0d92830e-serving-cert\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997491 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-etcd-serving-ca\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997519 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997542 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997569 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j22gj\" (UniqueName: \"kubernetes.io/projected/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-kube-api-access-j22gj\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997588 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g92d5\" (UniqueName: \"kubernetes.io/projected/b130afa1-783b-4b51-98bb-567c0d92830e-kube-api-access-g92d5\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997608 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l59t5\" (UniqueName: \"kubernetes.io/projected/dfe9c513-2c23-4583-937f-4a44fcf0452a-kube-api-access-l59t5\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997638 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838ccf5b-b653-42f4-be39-b6cc8f451440-config\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-serving-cert\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997679 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997703 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-image-import-ca\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997721 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-serving-cert\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997738 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-client\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997759 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-audit-dir\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997782 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shmgx\" (UniqueName: \"kubernetes.io/projected/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-kube-api-access-shmgx\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997815 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ec915743-c847-49f8-9d0e-c800b9fb2f2a-machine-approver-tls\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfe9c513-2c23-4583-937f-4a44fcf0452a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997858 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-audit\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997875 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-encryption-config\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997908 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/838ccf5b-b653-42f4-be39-b6cc8f451440-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997925 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-audit-policies\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997951 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-etcd-client\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.997991 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-config\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998011 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp498\" (UniqueName: \"kubernetes.io/projected/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-kube-api-access-tp498\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998030 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bpzs\" (UniqueName: \"kubernetes.io/projected/838ccf5b-b653-42f4-be39-b6cc8f451440-kube-api-access-7bpzs\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998048 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-serving-cert\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998068 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdmrn\" (UniqueName: \"kubernetes.io/projected/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-kube-api-access-cdmrn\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998084 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-service-ca\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998103 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-audit-policies\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998120 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e04b8c46-159a-4536-8f53-faff85690626-metrics-tls\") pod \"dns-operator-744455d44c-xxnvh\" (UID: \"e04b8c46-159a-4536-8f53-faff85690626\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998135 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-oauth-serving-cert\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998150 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-serving-cert\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998167 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998183 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-config\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998291 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-etcd-client\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998315 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfe9c513-2c23-4583-937f-4a44fcf0452a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998336 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2333c3b2-23cb-4e77-a88c-67f9118eb8de-serving-cert\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be880348-4fad-4b89-8e00-6dd336636f76-audit-dir\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998385 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2333c3b2-23cb-4e77-a88c-67f9118eb8de-config\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998402 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a93827e-e818-45a4-9276-bada4a616712-serving-cert\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998424 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-oauth-config\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998444 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kmkh\" (UniqueName: \"kubernetes.io/projected/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-kube-api-access-5kmkh\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998463 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998479 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgffx\" (UniqueName: \"kubernetes.io/projected/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-kube-api-access-cgffx\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998497 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbjw\" (UniqueName: \"kubernetes.io/projected/e04b8c46-159a-4536-8f53-faff85690626-kube-api-access-kqbjw\") pod \"dns-operator-744455d44c-xxnvh\" (UID: \"e04b8c46-159a-4536-8f53-faff85690626\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998521 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998539 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998557 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-client-ca\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998579 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-serving-cert\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998597 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngs4l\" (UniqueName: \"kubernetes.io/projected/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-kube-api-access-ngs4l\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998615 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998656 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dfe9c513-2c23-4583-937f-4a44fcf0452a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998674 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-audit-dir\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998693 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec915743-c847-49f8-9d0e-c800b9fb2f2a-auth-proxy-config\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998722 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlpf9\" (UniqueName: \"kubernetes.io/projected/604db804-d20c-4233-99d3-98690412d5cd-kube-api-access-hlpf9\") pod \"cluster-samples-operator-665b6dd947-7r26m\" (UID: \"604db804-d20c-4233-99d3-98690412d5cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998741 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2pxp\" (UniqueName: \"kubernetes.io/projected/7e2be296-6def-442a-8771-f9333e891d14-kube-api-access-z2pxp\") pod \"downloads-7954f5f757-rc25f\" (UID: \"7e2be296-6def-442a-8771-f9333e891d14\") " pod="openshift-console/downloads-7954f5f757-rc25f" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998761 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec915743-c847-49f8-9d0e-c800b9fb2f2a-config\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998780 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj56h\" (UniqueName: \"kubernetes.io/projected/2333c3b2-23cb-4e77-a88c-67f9118eb8de-kube-api-access-jj56h\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998797 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-images\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998804 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.998817 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:38 crc kubenswrapper[4960]: I1002 07:18:38.999532 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-ca\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.029502 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.034216 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.044328 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.046117 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.046241 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.047136 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.047536 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.047855 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.049549 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.049865 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.051777 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.055176 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.056241 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.057411 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.057718 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.057943 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.058060 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.065461 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.066011 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.068949 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dgl58"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.069271 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.069333 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.069291 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.070162 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.076294 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.079951 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.081629 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.082238 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.083520 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.084865 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nc4hj"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.085318 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.087177 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.087823 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.089172 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.090222 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.092000 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gzvxm"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.093221 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.094906 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.096043 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.096592 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.097669 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-96r9w"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100360 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-oauth-config\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100427 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kmkh\" (UniqueName: \"kubernetes.io/projected/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-kube-api-access-5kmkh\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100453 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100470 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgffx\" (UniqueName: \"kubernetes.io/projected/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-kube-api-access-cgffx\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbjw\" (UniqueName: \"kubernetes.io/projected/e04b8c46-159a-4536-8f53-faff85690626-kube-api-access-kqbjw\") pod \"dns-operator-744455d44c-xxnvh\" (UID: \"e04b8c46-159a-4536-8f53-faff85690626\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100524 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100541 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100574 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-client-ca\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100595 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-serving-cert\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100651 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngs4l\" (UniqueName: \"kubernetes.io/projected/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-kube-api-access-ngs4l\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100670 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100691 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dfe9c513-2c23-4583-937f-4a44fcf0452a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100710 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-audit-dir\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100746 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec915743-c847-49f8-9d0e-c800b9fb2f2a-auth-proxy-config\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100775 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlpf9\" (UniqueName: \"kubernetes.io/projected/604db804-d20c-4233-99d3-98690412d5cd-kube-api-access-hlpf9\") pod \"cluster-samples-operator-665b6dd947-7r26m\" (UID: \"604db804-d20c-4233-99d3-98690412d5cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100807 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2pxp\" (UniqueName: \"kubernetes.io/projected/7e2be296-6def-442a-8771-f9333e891d14-kube-api-access-z2pxp\") pod \"downloads-7954f5f757-rc25f\" (UID: \"7e2be296-6def-442a-8771-f9333e891d14\") " pod="openshift-console/downloads-7954f5f757-rc25f" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100825 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec915743-c847-49f8-9d0e-c800b9fb2f2a-config\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100846 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj56h\" (UniqueName: \"kubernetes.io/projected/2333c3b2-23cb-4e77-a88c-67f9118eb8de-kube-api-access-jj56h\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100891 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-images\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100930 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.100966 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-ca\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101001 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-service-ca-bundle\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101118 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101340 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-trusted-ca-bundle\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101386 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/604db804-d20c-4233-99d3-98690412d5cd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7r26m\" (UID: \"604db804-d20c-4233-99d3-98690412d5cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101403 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101422 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101465 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101485 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101523 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg6v7\" (UniqueName: \"kubernetes.io/projected/ec915743-c847-49f8-9d0e-c800b9fb2f2a-kube-api-access-pg6v7\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101542 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5dqd\" (UniqueName: \"kubernetes.io/projected/be880348-4fad-4b89-8e00-6dd336636f76-kube-api-access-l5dqd\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101559 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-config\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101599 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101618 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101634 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101655 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101698 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-client-ca\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101716 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2333c3b2-23cb-4e77-a88c-67f9118eb8de-trusted-ca\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101735 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfvlq\" (UniqueName: \"kubernetes.io/projected/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-kube-api-access-nfvlq\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101780 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7mfc\" (UniqueName: \"kubernetes.io/projected/7a93827e-e818-45a4-9276-bada4a616712-kube-api-access-z7mfc\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101807 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-encryption-config\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101860 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-node-pullsecrets\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101879 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-config\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101914 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-service-ca\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101950 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.101969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-serving-cert\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102042 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-config\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102061 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-config\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102078 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-config\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102115 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b130afa1-783b-4b51-98bb-567c0d92830e-serving-cert\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102132 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102145 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-etcd-serving-ca\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102164 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102203 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j22gj\" (UniqueName: \"kubernetes.io/projected/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-kube-api-access-j22gj\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102243 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g92d5\" (UniqueName: \"kubernetes.io/projected/b130afa1-783b-4b51-98bb-567c0d92830e-kube-api-access-g92d5\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102279 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l59t5\" (UniqueName: \"kubernetes.io/projected/dfe9c513-2c23-4583-937f-4a44fcf0452a-kube-api-access-l59t5\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102305 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838ccf5b-b653-42f4-be39-b6cc8f451440-config\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102343 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-serving-cert\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102363 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102382 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-image-import-ca\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102401 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-serving-cert\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102435 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-client\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102454 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-audit-dir\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102475 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shmgx\" (UniqueName: \"kubernetes.io/projected/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-kube-api-access-shmgx\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102530 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ec915743-c847-49f8-9d0e-c800b9fb2f2a-machine-approver-tls\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102549 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfe9c513-2c23-4583-937f-4a44fcf0452a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102567 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-audit\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102604 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102625 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-encryption-config\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102642 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/838ccf5b-b653-42f4-be39-b6cc8f451440-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102685 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-audit-policies\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102709 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-etcd-client\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102726 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-config\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102759 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp498\" (UniqueName: \"kubernetes.io/projected/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-kube-api-access-tp498\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102780 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bpzs\" (UniqueName: \"kubernetes.io/projected/838ccf5b-b653-42f4-be39-b6cc8f451440-kube-api-access-7bpzs\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102797 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-serving-cert\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102835 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdmrn\" (UniqueName: \"kubernetes.io/projected/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-kube-api-access-cdmrn\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-service-ca\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102875 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-audit-policies\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102887 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-audit-dir\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102911 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e04b8c46-159a-4536-8f53-faff85690626-metrics-tls\") pod \"dns-operator-744455d44c-xxnvh\" (UID: \"e04b8c46-159a-4536-8f53-faff85690626\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-oauth-serving-cert\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102929 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec915743-c847-49f8-9d0e-c800b9fb2f2a-config\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102950 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-serving-cert\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103004 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103025 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-config\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103052 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-etcd-client\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103086 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfe9c513-2c23-4583-937f-4a44fcf0452a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103111 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2333c3b2-23cb-4e77-a88c-67f9118eb8de-serving-cert\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103128 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be880348-4fad-4b89-8e00-6dd336636f76-audit-dir\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103169 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2333c3b2-23cb-4e77-a88c-67f9118eb8de-config\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.103190 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a93827e-e818-45a4-9276-bada4a616712-serving-cert\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.104260 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.104819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.105136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-images\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.105498 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.105801 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-audit-dir\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.102501 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.106812 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.106957 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838ccf5b-b653-42f4-be39-b6cc8f451440-config\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.107243 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.107478 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.108193 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-audit\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.108611 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec915743-c847-49f8-9d0e-c800b9fb2f2a-auth-proxy-config\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.109086 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-config\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.109425 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-image-import-ca\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.110054 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.111106 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-ca\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.111235 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-service-ca-bundle\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.111278 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-config\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.112341 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-audit-policies\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.112445 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-trusted-ca-bundle\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.112761 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a93827e-e818-45a4-9276-bada4a616712-serving-cert\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.115373 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.116648 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.116784 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.115368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-config\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.117520 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfe9c513-2c23-4583-937f-4a44fcf0452a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.117990 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.118370 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be880348-4fad-4b89-8e00-6dd336636f76-audit-dir\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.127192 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-oauth-serving-cert\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.127556 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-etcd-client\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.127827 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ec915743-c847-49f8-9d0e-c800b9fb2f2a-machine-approver-tls\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128059 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-etcd-serving-ca\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128088 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2333c3b2-23cb-4e77-a88c-67f9118eb8de-config\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128262 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-etcd-client\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128277 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-client-ca\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128354 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128475 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-serving-cert\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128622 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128773 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/838ccf5b-b653-42f4-be39-b6cc8f451440-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.128856 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-client\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.129156 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.129208 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.129543 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-serving-cert\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.129656 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.130201 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.130293 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2333c3b2-23cb-4e77-a88c-67f9118eb8de-serving-cert\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.130332 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b130afa1-783b-4b51-98bb-567c0d92830e-serving-cert\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.130631 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.131162 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.131391 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/604db804-d20c-4233-99d3-98690412d5cd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7r26m\" (UID: \"604db804-d20c-4233-99d3-98690412d5cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.131407 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-serving-cert\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.131531 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.131897 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-oauth-config\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.132282 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-encryption-config\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.132571 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.132831 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-client-ca\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.133695 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-audit-policies\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.134510 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-serving-cert\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.134514 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.135971 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a93827e-e818-45a4-9276-bada4a616712-config\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.137518 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-service-ca\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.137737 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-node-pullsecrets\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.138684 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.139773 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-etcd-service-ca\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.140655 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-config\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.142113 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.142413 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-config\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.143057 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.143287 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lddcs"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.143825 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-encryption-config\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.144774 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-serving-cert\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.144990 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.147241 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.154908 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.155898 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-89pc5"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.158733 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zcpbw"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.159567 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h8cqc"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.165694 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-serving-cert\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.165764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dfe9c513-2c23-4583-937f-4a44fcf0452a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.166451 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.174307 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2333c3b2-23cb-4e77-a88c-67f9118eb8de-trusted-ca\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.190187 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.190291 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.192617 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rc25f"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.196713 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vz5gr"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.198150 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.198856 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jzg6w"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.200518 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.201854 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.203116 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.204543 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nnbxf"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.206373 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-smfhg"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.206467 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.208039 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.209543 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fr4q7"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.211176 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.212867 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.213795 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xxnvh"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.215380 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.215501 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.217162 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8md6w"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.218433 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.219602 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.220852 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nc4hj"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.222496 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.223592 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gzvxm"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.224597 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.225777 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-96r9w"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.226756 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.227743 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.228742 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.229712 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dgl58"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.230673 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nnbxf"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.231639 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.232756 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.233739 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.234889 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-gpq9m"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.235528 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.236739 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-xkfg8"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.236960 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.239560 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gpq9m"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.239652 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.239729 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.240412 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xkfg8"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.241559 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bpfrh"] Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.242368 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.255633 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.262893 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b130afa1-783b-4b51-98bb-567c0d92830e-config\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.276955 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.287069 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e04b8c46-159a-4536-8f53-faff85690626-metrics-tls\") pod \"dns-operator-744455d44c-xxnvh\" (UID: \"e04b8c46-159a-4536-8f53-faff85690626\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.296340 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.315625 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.356142 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.377654 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.405283 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.416619 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.457237 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.477272 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.497334 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.517217 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.537141 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.556962 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.575916 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.596165 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.616270 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.636695 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.656133 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.675726 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.696687 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.716320 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.736052 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.756754 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.776215 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.796932 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.817353 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.836606 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.855791 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.876722 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.896481 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.916531 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.936706 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.956106 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.977099 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 07:18:39 crc kubenswrapper[4960]: I1002 07:18:39.997552 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.016877 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.035651 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.056307 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.075901 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.094299 4960 request.go:700] Waited for 1.008718602s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/configmaps?fieldSelector=metadata.name%3Dsigning-cabundle&limit=500&resourceVersion=0 Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.096352 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.115774 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.135858 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.155165 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.176815 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.195721 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.216832 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.236302 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.256443 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.277397 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.297279 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.317469 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.351821 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbjw\" (UniqueName: \"kubernetes.io/projected/e04b8c46-159a-4536-8f53-faff85690626-kube-api-access-kqbjw\") pod \"dns-operator-744455d44c-xxnvh\" (UID: \"e04b8c46-159a-4536-8f53-faff85690626\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.377152 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlpf9\" (UniqueName: \"kubernetes.io/projected/604db804-d20c-4233-99d3-98690412d5cd-kube-api-access-hlpf9\") pod \"cluster-samples-operator-665b6dd947-7r26m\" (UID: \"604db804-d20c-4233-99d3-98690412d5cd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.403006 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2pxp\" (UniqueName: \"kubernetes.io/projected/7e2be296-6def-442a-8771-f9333e891d14-kube-api-access-z2pxp\") pod \"downloads-7954f5f757-rc25f\" (UID: \"7e2be296-6def-442a-8771-f9333e891d14\") " pod="openshift-console/downloads-7954f5f757-rc25f" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.416253 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.417884 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kmkh\" (UniqueName: \"kubernetes.io/projected/4e0d9ffd-9187-4446-9a5b-87b1b999ff6e-kube-api-access-5kmkh\") pod \"machine-api-operator-5694c8668f-zcpbw\" (UID: \"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.437472 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.443944 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.448299 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.456575 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.485404 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.498866 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.550169 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgffx\" (UniqueName: \"kubernetes.io/projected/0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6-kube-api-access-cgffx\") pod \"apiserver-76f77b778f-h8cqc\" (UID: \"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6\") " pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.559341 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.559795 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj56h\" (UniqueName: \"kubernetes.io/projected/2333c3b2-23cb-4e77-a88c-67f9118eb8de-kube-api-access-jj56h\") pod \"console-operator-58897d9998-jzg6w\" (UID: \"2333c3b2-23cb-4e77-a88c-67f9118eb8de\") " pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.572517 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shmgx\" (UniqueName: \"kubernetes.io/projected/15752c4a-07e4-4fa6-9579-b5ac7ae26c12-kube-api-access-shmgx\") pod \"openshift-controller-manager-operator-756b6f6bc6-q2x4f\" (UID: \"15752c4a-07e4-4fa6-9579-b5ac7ae26c12\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.576542 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.584389 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.592337 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rc25f" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.597039 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.617393 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.638419 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.660373 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.676244 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.698013 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.712004 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.718254 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.720927 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.739425 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.773815 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp498\" (UniqueName: \"kubernetes.io/projected/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-kube-api-access-tp498\") pod \"controller-manager-879f6c89f-lddcs\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.778852 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m"] Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.792601 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bpzs\" (UniqueName: \"kubernetes.io/projected/838ccf5b-b653-42f4-be39-b6cc8f451440-kube-api-access-7bpzs\") pod \"openshift-apiserver-operator-796bbdcf4f-52fm9\" (UID: \"838ccf5b-b653-42f4-be39-b6cc8f451440\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.818720 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdmrn\" (UniqueName: \"kubernetes.io/projected/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-kube-api-access-cdmrn\") pod \"route-controller-manager-6576b87f9c-vcb9n\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.833556 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngs4l\" (UniqueName: \"kubernetes.io/projected/2d62d400-9f39-4905-a5a2-bfbfe3753dd1-kube-api-access-ngs4l\") pod \"apiserver-7bbb656c7d-2mgbj\" (UID: \"2d62d400-9f39-4905-a5a2-bfbfe3753dd1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.850384 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg6v7\" (UniqueName: \"kubernetes.io/projected/ec915743-c847-49f8-9d0e-c800b9fb2f2a-kube-api-access-pg6v7\") pod \"machine-approver-56656f9798-7sh7k\" (UID: \"ec915743-c847-49f8-9d0e-c800b9fb2f2a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.867104 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.870180 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfe9c513-2c23-4583-937f-4a44fcf0452a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.891537 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5dqd\" (UniqueName: \"kubernetes.io/projected/be880348-4fad-4b89-8e00-6dd336636f76-kube-api-access-l5dqd\") pod \"oauth-openshift-558db77b4-fr4q7\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.895528 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.904512 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.916585 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.924120 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jzg6w"] Oct 02 07:18:40 crc kubenswrapper[4960]: W1002 07:18:40.932734 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2333c3b2_23cb_4e77_a88c_67f9118eb8de.slice/crio-b00c24103aee2af5569431af9f8ba7d213b8595db7275a05ce388937068244a5 WatchSource:0}: Error finding container b00c24103aee2af5569431af9f8ba7d213b8595db7275a05ce388937068244a5: Status 404 returned error can't find the container with id b00c24103aee2af5569431af9f8ba7d213b8595db7275a05ce388937068244a5 Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.935539 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.960584 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f"] Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.981490 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7mfc\" (UniqueName: \"kubernetes.io/projected/7a93827e-e818-45a4-9276-bada4a616712-kube-api-access-z7mfc\") pod \"authentication-operator-69f744f599-89pc5\" (UID: \"7a93827e-e818-45a4-9276-bada4a616712\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.987038 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" Oct 02 07:18:40 crc kubenswrapper[4960]: I1002 07:18:40.999669 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfvlq\" (UniqueName: \"kubernetes.io/projected/ff8045d0-82c4-4054-8c26-f1bbf25dde9e-kube-api-access-nfvlq\") pod \"openshift-config-operator-7777fb866f-zbqtp\" (UID: \"ff8045d0-82c4-4054-8c26-f1bbf25dde9e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.002591 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.002902 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:41 crc kubenswrapper[4960]: W1002 07:18:41.004898 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15752c4a_07e4_4fa6_9579_b5ac7ae26c12.slice/crio-ae04b93580572b010eea1fdaf2a36f981e3ed975506279114847d3a55f964663 WatchSource:0}: Error finding container ae04b93580572b010eea1fdaf2a36f981e3ed975506279114847d3a55f964663: Status 404 returned error can't find the container with id ae04b93580572b010eea1fdaf2a36f981e3ed975506279114847d3a55f964663 Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.004909 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zcpbw"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.007795 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xxnvh"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.014004 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g92d5\" (UniqueName: \"kubernetes.io/projected/b130afa1-783b-4b51-98bb-567c0d92830e-kube-api-access-g92d5\") pod \"etcd-operator-b45778765-smfhg\" (UID: \"b130afa1-783b-4b51-98bb-567c0d92830e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.016994 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.028296 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.032448 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j22gj\" (UniqueName: \"kubernetes.io/projected/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-kube-api-access-j22gj\") pod \"console-f9d7485db-vz5gr\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:41 crc kubenswrapper[4960]: W1002 07:18:41.039182 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode04b8c46_159a_4536_8f53_faff85690626.slice/crio-975ab9052f0296115e939c225d896b2657449640ad1c30aabd9b02f5c3319f6d WatchSource:0}: Error finding container 975ab9052f0296115e939c225d896b2657449640ad1c30aabd9b02f5c3319f6d: Status 404 returned error can't find the container with id 975ab9052f0296115e939c225d896b2657449640ad1c30aabd9b02f5c3319f6d Oct 02 07:18:41 crc kubenswrapper[4960]: W1002 07:18:41.047671 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e0d9ffd_9187_4446_9a5b_87b1b999ff6e.slice/crio-20d467cdd590d27b7b9f5de4bf829375e92d84241c560dac1f8ce6695b68157a WatchSource:0}: Error finding container 20d467cdd590d27b7b9f5de4bf829375e92d84241c560dac1f8ce6695b68157a: Status 404 returned error can't find the container with id 20d467cdd590d27b7b9f5de4bf829375e92d84241c560dac1f8ce6695b68157a Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.049730 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h8cqc"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.050758 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rc25f"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.055684 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l59t5\" (UniqueName: \"kubernetes.io/projected/dfe9c513-2c23-4583-937f-4a44fcf0452a-kube-api-access-l59t5\") pod \"cluster-image-registry-operator-dc59b4c8b-9vwr2\" (UID: \"dfe9c513-2c23-4583-937f-4a44fcf0452a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.057131 4960 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.076179 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.087863 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.091131 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lddcs"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.094621 4960 request.go:700] Waited for 1.88786481s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.096494 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.118116 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.143717 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.156246 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.162514 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.165222 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" event={"ID":"e04b8c46-159a-4536-8f53-faff85690626","Type":"ContainerStarted","Data":"975ab9052f0296115e939c225d896b2657449640ad1c30aabd9b02f5c3319f6d"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.169032 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" event={"ID":"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e","Type":"ContainerStarted","Data":"20d467cdd590d27b7b9f5de4bf829375e92d84241c560dac1f8ce6695b68157a"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.169261 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.176034 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.177141 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" event={"ID":"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6","Type":"ContainerStarted","Data":"b14a7e71b802211e0c3e5af0882db29f7404f056e1e708491c2024ad7ab30ec9"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.180136 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" event={"ID":"2333c3b2-23cb-4e77-a88c-67f9118eb8de","Type":"ContainerStarted","Data":"7f639dd433a97124bb4ca3a20e4962e2a1ad85cc6516609020f1c6aeb15f4863"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.180169 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" event={"ID":"2333c3b2-23cb-4e77-a88c-67f9118eb8de","Type":"ContainerStarted","Data":"b00c24103aee2af5569431af9f8ba7d213b8595db7275a05ce388937068244a5"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.183270 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.189241 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" event={"ID":"15752c4a-07e4-4fa6-9579-b5ac7ae26c12","Type":"ContainerStarted","Data":"ae04b93580572b010eea1fdaf2a36f981e3ed975506279114847d3a55f964663"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.189304 4960 patch_prober.go:28] interesting pod/console-operator-58897d9998-jzg6w container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.189361 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" podUID="2333c3b2-23cb-4e77-a88c-67f9118eb8de" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.191517 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rc25f" event={"ID":"7e2be296-6def-442a-8771-f9333e891d14","Type":"ContainerStarted","Data":"c8c551986783258479770a012e340c473547e684492bc8c4414e8013bc20034d"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.195005 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" event={"ID":"604db804-d20c-4233-99d3-98690412d5cd","Type":"ContainerStarted","Data":"076084d72b6b2b2b724b81aa34a21fa44deee88d1e571af7dc840f14a8d2645e"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.195079 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" event={"ID":"604db804-d20c-4233-99d3-98690412d5cd","Type":"ContainerStarted","Data":"a97580777b34f5bd62aa12be0511e0b6491f0c32eed4158ebbe9be43ab4814a1"} Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.199830 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.218284 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.237899 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.273473 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.277505 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.295131 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.299743 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.322356 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335437 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335578 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-certificates\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335597 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-tls\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335633 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-bound-sa-token\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-trusted-ca\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335693 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hqlf\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-kube-api-access-2hqlf\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335716 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.335739 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: E1002 07:18:41.336074 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:41.836058586 +0000 UTC m=+142.868004943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.436890 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437153 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4ffa2a8-76aa-48e8-a617-597c63719217-trusted-ca\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437193 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-certificates\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437218 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf5h8\" (UniqueName: \"kubernetes.io/projected/08abc337-7b21-4bc4-8e31-7b594d654312-kube-api-access-qf5h8\") pod \"migrator-59844c95c7-9mrbf\" (UID: \"08abc337-7b21-4bc4-8e31-7b594d654312\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437259 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdzl2\" (UniqueName: \"kubernetes.io/projected/dffe37b6-aae9-49ff-a8a7-2c64360d748b-kube-api-access-pdzl2\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437316 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4ffa2a8-76aa-48e8-a617-597c63719217-metrics-tls\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437331 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7slb\" (UniqueName: \"kubernetes.io/projected/c3480a6e-4787-4d30-b05b-da16f1356be4-kube-api-access-f7slb\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437356 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-plugins-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437370 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c3480a6e-4787-4d30-b05b-da16f1356be4-srv-cert\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437392 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be65bb8e-a260-43f7-8e00-78c4158a683e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437409 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3d80dacd-d5ee-4616-878d-0563d69b6b6f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4xjvf\" (UID: \"3d80dacd-d5ee-4616-878d-0563d69b6b6f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437427 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twmrv\" (UniqueName: \"kubernetes.io/projected/f58114fb-9f32-45ab-9a99-5c4a378353de-kube-api-access-twmrv\") pod \"ingress-canary-gpq9m\" (UID: \"f58114fb-9f32-45ab-9a99-5c4a378353de\") " pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437457 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dffe37b6-aae9-49ff-a8a7-2c64360d748b-images\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437472 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8vjf\" (UniqueName: \"kubernetes.io/projected/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-kube-api-access-x8vjf\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437494 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hsbf5\" (UID: \"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437511 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97a0f863-045c-4251-865d-327d1c1397e5-secret-volume\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437526 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c019958e-ed31-4dc5-b8f7-135cf1e69f70-config\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437544 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hqlf\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-kube-api-access-2hqlf\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437559 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn9tx\" (UniqueName: \"kubernetes.io/projected/7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34-kube-api-access-wn9tx\") pod \"package-server-manager-789f6589d5-hsbf5\" (UID: \"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437588 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1070796-a85a-4cd5-a302-bdaa2541db68-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437604 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97a0f863-045c-4251-865d-327d1c1397e5-config-volume\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437618 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-default-certificate\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437663 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69def292-5958-44c2-a8fb-362bbdb897a0-tmpfs\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437702 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e34b1da7-5b94-4313-81f8-401e1a798a09-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437720 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437753 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926160e-8baa-4e5b-b0ce-2de658e7250f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437769 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-stats-auth\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437809 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmnm\" (UniqueName: \"kubernetes.io/projected/b4ffa2a8-76aa-48e8-a617-597c63719217-kube-api-access-cvmnm\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437824 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5fnc\" (UniqueName: \"kubernetes.io/projected/fef811a3-779c-4bf7-8eb2-cbab93f43f8e-kube-api-access-f5fnc\") pod \"multus-admission-controller-857f4d67dd-gzvxm\" (UID: \"fef811a3-779c-4bf7-8eb2-cbab93f43f8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437856 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69def292-5958-44c2-a8fb-362bbdb897a0-apiservice-cert\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.437903 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3cda2a13-5fea-4023-afb5-ed377f21b98e-signing-cabundle\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: E1002 07:18:41.438000 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:41.937943835 +0000 UTC m=+142.969890182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438089 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5926160e-8baa-4e5b-b0ce-2de658e7250f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438193 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-metrics-certs\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438242 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-metrics-tls\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438281 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dffe37b6-aae9-49ff-a8a7-2c64360d748b-proxy-tls\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438327 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfswd\" (UniqueName: \"kubernetes.io/projected/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-kube-api-access-kfswd\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438371 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-socket-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438488 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57150513-840e-47ef-8e38-91b3b090a380-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnbvq\" (UniqueName: \"kubernetes.io/projected/e919b925-67b0-475f-b01c-e37ae52c3dbb-kube-api-access-gnbvq\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438620 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438667 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-csi-data-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438692 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-node-bootstrap-token\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438760 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-tls\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438807 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6b93888-4d26-4500-ae07-aefff1dc0a7b-service-ca-bundle\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438833 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438890 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5926160e-8baa-4e5b-b0ce-2de658e7250f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.438962 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd9dz\" (UniqueName: \"kubernetes.io/projected/0122bd5f-ff81-4468-aac9-32f277f12e68-kube-api-access-cd9dz\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.439598 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e919b925-67b0-475f-b01c-e37ae52c3dbb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440065 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c019958e-ed31-4dc5-b8f7-135cf1e69f70-serving-cert\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440104 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-bound-sa-token\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440120 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be65bb8e-a260-43f7-8e00-78c4158a683e-srv-cert\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440145 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-trusted-ca\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440163 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-certs\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440219 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-config-volume\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4ffa2a8-76aa-48e8-a617-597c63719217-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440253 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c3480a6e-4787-4d30-b05b-da16f1356be4-profile-collector-cert\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440282 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-registration-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440322 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34b1da7-5b94-4313-81f8-401e1a798a09-config\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440349 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1070796-a85a-4cd5-a302-bdaa2541db68-config\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.440397 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fk7r\" (UniqueName: \"kubernetes.io/projected/69def292-5958-44c2-a8fb-362bbdb897a0-kube-api-access-8fk7r\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442266 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srr2d\" (UniqueName: \"kubernetes.io/projected/3cda2a13-5fea-4023-afb5-ed377f21b98e-kube-api-access-srr2d\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442295 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e34b1da7-5b94-4313-81f8-401e1a798a09-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442342 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57150513-840e-47ef-8e38-91b3b090a380-proxy-tls\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442358 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3cda2a13-5fea-4023-afb5-ed377f21b98e-signing-key\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442375 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442413 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djbnt\" (UniqueName: \"kubernetes.io/projected/97a0f863-045c-4251-865d-327d1c1397e5-kube-api-access-djbnt\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442451 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69def292-5958-44c2-a8fb-362bbdb897a0-webhook-cert\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442467 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fef811a3-779c-4bf7-8eb2-cbab93f43f8e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gzvxm\" (UID: \"fef811a3-779c-4bf7-8eb2-cbab93f43f8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442483 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9xtc\" (UniqueName: \"kubernetes.io/projected/57150513-840e-47ef-8e38-91b3b090a380-kube-api-access-n9xtc\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442499 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-mountpoint-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442539 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns8gv\" (UniqueName: \"kubernetes.io/projected/be65bb8e-a260-43f7-8e00-78c4158a683e-kube-api-access-ns8gv\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442555 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7x24\" (UniqueName: \"kubernetes.io/projected/f6b93888-4d26-4500-ae07-aefff1dc0a7b-kube-api-access-h7x24\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442579 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f58114fb-9f32-45ab-9a99-5c4a378353de-cert\") pod \"ingress-canary-gpq9m\" (UID: \"f58114fb-9f32-45ab-9a99-5c4a378353de\") " pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442595 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75tmp\" (UniqueName: \"kubernetes.io/projected/3d80dacd-d5ee-4616-878d-0563d69b6b6f-kube-api-access-75tmp\") pod \"control-plane-machine-set-operator-78cbb6b69f-4xjvf\" (UID: \"3d80dacd-d5ee-4616-878d-0563d69b6b6f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442620 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nspwb\" (UniqueName: \"kubernetes.io/projected/c019958e-ed31-4dc5-b8f7-135cf1e69f70-kube-api-access-nspwb\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442635 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxr4c\" (UniqueName: \"kubernetes.io/projected/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-kube-api-access-dxr4c\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442670 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dffe37b6-aae9-49ff-a8a7-2c64360d748b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442685 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1070796-a85a-4cd5-a302-bdaa2541db68-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.442728 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e919b925-67b0-475f-b01c-e37ae52c3dbb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.448085 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.453863 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-trusted-ca\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.461227 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-certificates\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.466903 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.467567 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-tls\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.467872 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.493770 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hqlf\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-kube-api-access-2hqlf\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.515898 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-bound-sa-token\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.543774 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-stats-auth\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.543823 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926160e-8baa-4e5b-b0ce-2de658e7250f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.543860 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmnm\" (UniqueName: \"kubernetes.io/projected/b4ffa2a8-76aa-48e8-a617-597c63719217-kube-api-access-cvmnm\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.543884 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5fnc\" (UniqueName: \"kubernetes.io/projected/fef811a3-779c-4bf7-8eb2-cbab93f43f8e-kube-api-access-f5fnc\") pod \"multus-admission-controller-857f4d67dd-gzvxm\" (UID: \"fef811a3-779c-4bf7-8eb2-cbab93f43f8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.543906 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69def292-5958-44c2-a8fb-362bbdb897a0-apiservice-cert\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.543929 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3cda2a13-5fea-4023-afb5-ed377f21b98e-signing-cabundle\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.543958 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5926160e-8baa-4e5b-b0ce-2de658e7250f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544000 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-metrics-tls\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544024 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-metrics-certs\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544046 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dffe37b6-aae9-49ff-a8a7-2c64360d748b-proxy-tls\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544066 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfswd\" (UniqueName: \"kubernetes.io/projected/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-kube-api-access-kfswd\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544088 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-socket-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544114 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57150513-840e-47ef-8e38-91b3b090a380-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544137 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnbvq\" (UniqueName: \"kubernetes.io/projected/e919b925-67b0-475f-b01c-e37ae52c3dbb-kube-api-access-gnbvq\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544160 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-csi-data-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544188 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-node-bootstrap-token\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544211 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5926160e-8baa-4e5b-b0ce-2de658e7250f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544236 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6b93888-4d26-4500-ae07-aefff1dc0a7b-service-ca-bundle\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544258 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544292 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd9dz\" (UniqueName: \"kubernetes.io/projected/0122bd5f-ff81-4468-aac9-32f277f12e68-kube-api-access-cd9dz\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544314 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c019958e-ed31-4dc5-b8f7-135cf1e69f70-serving-cert\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544334 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e919b925-67b0-475f-b01c-e37ae52c3dbb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544354 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be65bb8e-a260-43f7-8e00-78c4158a683e-srv-cert\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544373 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-certs\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544392 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4ffa2a8-76aa-48e8-a617-597c63719217-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544413 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c3480a6e-4787-4d30-b05b-da16f1356be4-profile-collector-cert\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544438 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-config-volume\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544459 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-registration-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544492 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34b1da7-5b94-4313-81f8-401e1a798a09-config\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544515 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1070796-a85a-4cd5-a302-bdaa2541db68-config\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544537 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fk7r\" (UniqueName: \"kubernetes.io/projected/69def292-5958-44c2-a8fb-362bbdb897a0-kube-api-access-8fk7r\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544559 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srr2d\" (UniqueName: \"kubernetes.io/projected/3cda2a13-5fea-4023-afb5-ed377f21b98e-kube-api-access-srr2d\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544581 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e34b1da7-5b94-4313-81f8-401e1a798a09-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544601 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3cda2a13-5fea-4023-afb5-ed377f21b98e-signing-key\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544621 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57150513-840e-47ef-8e38-91b3b090a380-proxy-tls\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544640 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544665 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djbnt\" (UniqueName: \"kubernetes.io/projected/97a0f863-045c-4251-865d-327d1c1397e5-kube-api-access-djbnt\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544687 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69def292-5958-44c2-a8fb-362bbdb897a0-webhook-cert\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544712 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-mountpoint-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544736 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fef811a3-779c-4bf7-8eb2-cbab93f43f8e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gzvxm\" (UID: \"fef811a3-779c-4bf7-8eb2-cbab93f43f8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544756 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9xtc\" (UniqueName: \"kubernetes.io/projected/57150513-840e-47ef-8e38-91b3b090a380-kube-api-access-n9xtc\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544775 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns8gv\" (UniqueName: \"kubernetes.io/projected/be65bb8e-a260-43f7-8e00-78c4158a683e-kube-api-access-ns8gv\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544791 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75tmp\" (UniqueName: \"kubernetes.io/projected/3d80dacd-d5ee-4616-878d-0563d69b6b6f-kube-api-access-75tmp\") pod \"control-plane-machine-set-operator-78cbb6b69f-4xjvf\" (UID: \"3d80dacd-d5ee-4616-878d-0563d69b6b6f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544805 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7x24\" (UniqueName: \"kubernetes.io/projected/f6b93888-4d26-4500-ae07-aefff1dc0a7b-kube-api-access-h7x24\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544819 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f58114fb-9f32-45ab-9a99-5c4a378353de-cert\") pod \"ingress-canary-gpq9m\" (UID: \"f58114fb-9f32-45ab-9a99-5c4a378353de\") " pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544835 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nspwb\" (UniqueName: \"kubernetes.io/projected/c019958e-ed31-4dc5-b8f7-135cf1e69f70-kube-api-access-nspwb\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544852 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxr4c\" (UniqueName: \"kubernetes.io/projected/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-kube-api-access-dxr4c\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544884 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dffe37b6-aae9-49ff-a8a7-2c64360d748b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544901 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1070796-a85a-4cd5-a302-bdaa2541db68-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544927 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e919b925-67b0-475f-b01c-e37ae52c3dbb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.544953 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4ffa2a8-76aa-48e8-a617-597c63719217-trusted-ca\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545034 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf5h8\" (UniqueName: \"kubernetes.io/projected/08abc337-7b21-4bc4-8e31-7b594d654312-kube-api-access-qf5h8\") pod \"migrator-59844c95c7-9mrbf\" (UID: \"08abc337-7b21-4bc4-8e31-7b594d654312\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545057 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdzl2\" (UniqueName: \"kubernetes.io/projected/dffe37b6-aae9-49ff-a8a7-2c64360d748b-kube-api-access-pdzl2\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545087 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4ffa2a8-76aa-48e8-a617-597c63719217-metrics-tls\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545109 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7slb\" (UniqueName: \"kubernetes.io/projected/c3480a6e-4787-4d30-b05b-da16f1356be4-kube-api-access-f7slb\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545124 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-plugins-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545138 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c3480a6e-4787-4d30-b05b-da16f1356be4-srv-cert\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545154 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be65bb8e-a260-43f7-8e00-78c4158a683e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545177 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3d80dacd-d5ee-4616-878d-0563d69b6b6f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4xjvf\" (UID: \"3d80dacd-d5ee-4616-878d-0563d69b6b6f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545202 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twmrv\" (UniqueName: \"kubernetes.io/projected/f58114fb-9f32-45ab-9a99-5c4a378353de-kube-api-access-twmrv\") pod \"ingress-canary-gpq9m\" (UID: \"f58114fb-9f32-45ab-9a99-5c4a378353de\") " pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545219 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dffe37b6-aae9-49ff-a8a7-2c64360d748b-images\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545234 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8vjf\" (UniqueName: \"kubernetes.io/projected/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-kube-api-access-x8vjf\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545252 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hsbf5\" (UID: \"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545268 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn9tx\" (UniqueName: \"kubernetes.io/projected/7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34-kube-api-access-wn9tx\") pod \"package-server-manager-789f6589d5-hsbf5\" (UID: \"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545283 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97a0f863-045c-4251-865d-327d1c1397e5-secret-volume\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.545297 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c019958e-ed31-4dc5-b8f7-135cf1e69f70-config\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.546650 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.546681 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1070796-a85a-4cd5-a302-bdaa2541db68-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.546699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97a0f863-045c-4251-865d-327d1c1397e5-config-volume\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.546717 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-default-certificate\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.546734 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69def292-5958-44c2-a8fb-362bbdb897a0-tmpfs\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.546751 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e34b1da7-5b94-4313-81f8-401e1a798a09-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.550742 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-config-volume\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.552546 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.557696 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6b93888-4d26-4500-ae07-aefff1dc0a7b-service-ca-bundle\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.558325 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-mountpoint-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.559537 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-registration-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.560282 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34b1da7-5b94-4313-81f8-401e1a798a09-config\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.560697 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1070796-a85a-4cd5-a302-bdaa2541db68-config\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.561080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926160e-8baa-4e5b-b0ce-2de658e7250f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.561486 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3cda2a13-5fea-4023-afb5-ed377f21b98e-signing-cabundle\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.561897 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dffe37b6-aae9-49ff-a8a7-2c64360d748b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.562070 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e34b1da7-5b94-4313-81f8-401e1a798a09-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.562593 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e919b925-67b0-475f-b01c-e37ae52c3dbb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.566309 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57150513-840e-47ef-8e38-91b3b090a380-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.566382 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-socket-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.566557 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4ffa2a8-76aa-48e8-a617-597c63719217-trusted-ca\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.566644 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-plugins-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.567156 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-csi-data-dir\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.567575 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c019958e-ed31-4dc5-b8f7-135cf1e69f70-config\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.567619 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/69def292-5958-44c2-a8fb-362bbdb897a0-tmpfs\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.567933 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97a0f863-045c-4251-865d-327d1c1397e5-config-volume\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.568358 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dffe37b6-aae9-49ff-a8a7-2c64360d748b-images\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: E1002 07:18:41.568393 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.068373378 +0000 UTC m=+143.100319665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.572257 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e919b925-67b0-475f-b01c-e37ae52c3dbb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.574560 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be65bb8e-a260-43f7-8e00-78c4158a683e-srv-cert\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.574928 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/69def292-5958-44c2-a8fb-362bbdb897a0-webhook-cert\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.578588 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.578687 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c019958e-ed31-4dc5-b8f7-135cf1e69f70-serving-cert\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.590239 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-89pc5"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.592508 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/69def292-5958-44c2-a8fb-362bbdb897a0-apiservice-cert\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.592529 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-metrics-certs\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.593689 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-certs\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.595295 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4ffa2a8-76aa-48e8-a617-597c63719217-metrics-tls\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.596315 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dffe37b6-aae9-49ff-a8a7-2c64360d748b-proxy-tls\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.596352 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fef811a3-779c-4bf7-8eb2-cbab93f43f8e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gzvxm\" (UID: \"fef811a3-779c-4bf7-8eb2-cbab93f43f8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.596891 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3cda2a13-5fea-4023-afb5-ed377f21b98e-signing-key\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.597653 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djbnt\" (UniqueName: \"kubernetes.io/projected/97a0f863-045c-4251-865d-327d1c1397e5-kube-api-access-djbnt\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.598598 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c3480a6e-4787-4d30-b05b-da16f1356be4-profile-collector-cert\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.599107 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-stats-auth\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.600500 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c3480a6e-4787-4d30-b05b-da16f1356be4-srv-cert\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.601323 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be65bb8e-a260-43f7-8e00-78c4158a683e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.607352 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3d80dacd-d5ee-4616-878d-0563d69b6b6f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4xjvf\" (UID: \"3d80dacd-d5ee-4616-878d-0563d69b6b6f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.609454 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97a0f863-045c-4251-865d-327d1c1397e5-secret-volume\") pod \"collect-profiles-29323155-g6jvn\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.609517 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f6b93888-4d26-4500-ae07-aefff1dc0a7b-default-certificate\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.609563 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57150513-840e-47ef-8e38-91b3b090a380-proxy-tls\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.609635 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hsbf5\" (UID: \"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.609792 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f58114fb-9f32-45ab-9a99-5c4a378353de-cert\") pod \"ingress-canary-gpq9m\" (UID: \"f58114fb-9f32-45ab-9a99-5c4a378353de\") " pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.610031 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-metrics-tls\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.610285 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-node-bootstrap-token\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.617054 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5926160e-8baa-4e5b-b0ce-2de658e7250f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.619591 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1070796-a85a-4cd5-a302-bdaa2541db68-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.630019 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4ffa2a8-76aa-48e8-a617-597c63719217-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.644692 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd9dz\" (UniqueName: \"kubernetes.io/projected/0122bd5f-ff81-4468-aac9-32f277f12e68-kube-api-access-cd9dz\") pod \"marketplace-operator-79b997595-96r9w\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.646677 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fr4q7"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.647240 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:41 crc kubenswrapper[4960]: E1002 07:18:41.647825 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.147805661 +0000 UTC m=+143.179751948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.653447 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fk7r\" (UniqueName: \"kubernetes.io/projected/69def292-5958-44c2-a8fb-362bbdb897a0-kube-api-access-8fk7r\") pod \"packageserver-d55dfcdfc-wdkwq\" (UID: \"69def292-5958-44c2-a8fb-362bbdb897a0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.672679 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.680955 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e34b1da7-5b94-4313-81f8-401e1a798a09-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z8gjv\" (UID: \"e34b1da7-5b94-4313-81f8-401e1a798a09\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.699641 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.721382 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9xtc\" (UniqueName: \"kubernetes.io/projected/57150513-840e-47ef-8e38-91b3b090a380-kube-api-access-n9xtc\") pod \"machine-config-controller-84d6567774-xkvwr\" (UID: \"57150513-840e-47ef-8e38-91b3b090a380\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.723589 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.729175 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srr2d\" (UniqueName: \"kubernetes.io/projected/3cda2a13-5fea-4023-afb5-ed377f21b98e-kube-api-access-srr2d\") pod \"service-ca-9c57cc56f-nc4hj\" (UID: \"3cda2a13-5fea-4023-afb5-ed377f21b98e\") " pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.731143 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.732677 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-smfhg"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.736682 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns8gv\" (UniqueName: \"kubernetes.io/projected/be65bb8e-a260-43f7-8e00-78c4158a683e-kube-api-access-ns8gv\") pod \"olm-operator-6b444d44fb-2h9xd\" (UID: \"be65bb8e-a260-43f7-8e00-78c4158a683e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.741252 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.753939 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: E1002 07:18:41.754436 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.254417503 +0000 UTC m=+143.286363790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.766799 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.776131 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75tmp\" (UniqueName: \"kubernetes.io/projected/3d80dacd-d5ee-4616-878d-0563d69b6b6f-kube-api-access-75tmp\") pod \"control-plane-machine-set-operator-78cbb6b69f-4xjvf\" (UID: \"3d80dacd-d5ee-4616-878d-0563d69b6b6f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.777571 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.782822 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7x24\" (UniqueName: \"kubernetes.io/projected/f6b93888-4d26-4500-ae07-aefff1dc0a7b-kube-api-access-h7x24\") pod \"router-default-5444994796-9lf9c\" (UID: \"f6b93888-4d26-4500-ae07-aefff1dc0a7b\") " pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.789479 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" Oct 02 07:18:41 crc kubenswrapper[4960]: W1002 07:18:41.802820 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d62d400_9f39_4905_a5a2_bfbfe3753dd1.slice/crio-56fd05328cadfdd648b483efb2868cec67da8135be2d7d3b90997ad63f1ae818 WatchSource:0}: Error finding container 56fd05328cadfdd648b483efb2868cec67da8135be2d7d3b90997ad63f1ae818: Status 404 returned error can't find the container with id 56fd05328cadfdd648b483efb2868cec67da8135be2d7d3b90997ad63f1ae818 Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.803895 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmnm\" (UniqueName: \"kubernetes.io/projected/b4ffa2a8-76aa-48e8-a617-597c63719217-kube-api-access-cvmnm\") pod \"ingress-operator-5b745b69d9-9r8xz\" (UID: \"b4ffa2a8-76aa-48e8-a617-597c63719217\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.829148 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5926160e-8baa-4e5b-b0ce-2de658e7250f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sznw8\" (UID: \"5926160e-8baa-4e5b-b0ce-2de658e7250f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.838799 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5fnc\" (UniqueName: \"kubernetes.io/projected/fef811a3-779c-4bf7-8eb2-cbab93f43f8e-kube-api-access-f5fnc\") pod \"multus-admission-controller-857f4d67dd-gzvxm\" (UID: \"fef811a3-779c-4bf7-8eb2-cbab93f43f8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.844973 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vz5gr"] Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.855593 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf5h8\" (UniqueName: \"kubernetes.io/projected/08abc337-7b21-4bc4-8e31-7b594d654312-kube-api-access-qf5h8\") pod \"migrator-59844c95c7-9mrbf\" (UID: \"08abc337-7b21-4bc4-8e31-7b594d654312\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.856158 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:41 crc kubenswrapper[4960]: E1002 07:18:41.856996 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.356958363 +0000 UTC m=+143.388904650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.873741 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7slb\" (UniqueName: \"kubernetes.io/projected/c3480a6e-4787-4d30-b05b-da16f1356be4-kube-api-access-f7slb\") pod \"catalog-operator-68c6474976-qcf8z\" (UID: \"c3480a6e-4787-4d30-b05b-da16f1356be4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.895516 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nspwb\" (UniqueName: \"kubernetes.io/projected/c019958e-ed31-4dc5-b8f7-135cf1e69f70-kube-api-access-nspwb\") pod \"service-ca-operator-777779d784-dgl58\" (UID: \"c019958e-ed31-4dc5-b8f7-135cf1e69f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.916434 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdzl2\" (UniqueName: \"kubernetes.io/projected/dffe37b6-aae9-49ff-a8a7-2c64360d748b-kube-api-access-pdzl2\") pod \"machine-config-operator-74547568cd-9lz9x\" (UID: \"dffe37b6-aae9-49ff-a8a7-2c64360d748b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:41 crc kubenswrapper[4960]: W1002 07:18:41.919567 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5911cf0_70ba_44a3_8c49_4cd66cca73b7.slice/crio-6ef0935cde1ec223e9c4d5726645648e4cf73f5c2fbd2fd4fdffcd8405bd564f WatchSource:0}: Error finding container 6ef0935cde1ec223e9c4d5726645648e4cf73f5c2fbd2fd4fdffcd8405bd564f: Status 404 returned error can't find the container with id 6ef0935cde1ec223e9c4d5726645648e4cf73f5c2fbd2fd4fdffcd8405bd564f Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.939909 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxr4c\" (UniqueName: \"kubernetes.io/projected/27d5c5ed-4a15-4889-9d0f-8fca8ba4f932-kube-api-access-dxr4c\") pod \"dns-default-xkfg8\" (UID: \"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932\") " pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.951126 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.956571 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1070796-a85a-4cd5-a302-bdaa2541db68-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-h6vf5\" (UID: \"d1070796-a85a-4cd5-a302-bdaa2541db68\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.957489 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.958380 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:41 crc kubenswrapper[4960]: E1002 07:18:41.958745 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.458721368 +0000 UTC m=+143.490667655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.965932 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.973096 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.984076 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfswd\" (UniqueName: \"kubernetes.io/projected/f1b94355-cfe1-447e-b2fe-dcc1730d7e39-kube-api-access-kfswd\") pod \"machine-config-server-bpfrh\" (UID: \"f1b94355-cfe1-447e-b2fe-dcc1730d7e39\") " pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.984455 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:41 crc kubenswrapper[4960]: I1002 07:18:41.991649 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnbvq\" (UniqueName: \"kubernetes.io/projected/e919b925-67b0-475f-b01c-e37ae52c3dbb-kube-api-access-gnbvq\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmx7\" (UID: \"e919b925-67b0-475f-b01c-e37ae52c3dbb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.005681 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.010574 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.021617 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twmrv\" (UniqueName: \"kubernetes.io/projected/f58114fb-9f32-45ab-9a99-5c4a378353de-kube-api-access-twmrv\") pod \"ingress-canary-gpq9m\" (UID: \"f58114fb-9f32-45ab-9a99-5c4a378353de\") " pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.031462 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8vjf\" (UniqueName: \"kubernetes.io/projected/d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327-kube-api-access-x8vjf\") pod \"csi-hostpathplugin-nnbxf\" (UID: \"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327\") " pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.038903 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.045393 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.052271 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.060625 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.065368 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.565345681 +0000 UTC m=+143.597291968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.066807 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.070296 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn9tx\" (UniqueName: \"kubernetes.io/projected/7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34-kube-api-access-wn9tx\") pod \"package-server-manager-789f6589d5-hsbf5\" (UID: \"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.084589 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.098486 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.118846 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.126522 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gpq9m" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.133606 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.135589 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq"] Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.145074 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bpfrh" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.166840 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.167469 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.667452727 +0000 UTC m=+143.699399014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.237817 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" event={"ID":"b130afa1-783b-4b51-98bb-567c0d92830e","Type":"ContainerStarted","Data":"625538bda0a28734a3cf59b1cf34aec9384d6dd7d5223886a0fedead22db0dcf"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.247906 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" event={"ID":"7a93827e-e818-45a4-9276-bada4a616712","Type":"ContainerStarted","Data":"7d97017a1d7c8ffca4c2ad7100f7d8d0f1908899b36cda91f7014be5217150a7"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.247959 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" event={"ID":"7a93827e-e818-45a4-9276-bada4a616712","Type":"ContainerStarted","Data":"a73e1d72412c76c845d0b5602cab8c8312a0628b0f6f77273d9f0e24d961b98c"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.250538 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv"] Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.254010 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vz5gr" event={"ID":"a5911cf0-70ba-44a3-8c49-4cd66cca73b7","Type":"ContainerStarted","Data":"6ef0935cde1ec223e9c4d5726645648e4cf73f5c2fbd2fd4fdffcd8405bd564f"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.256013 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn"] Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.270270 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.270628 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.770610877 +0000 UTC m=+143.802557164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.276279 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" event={"ID":"604db804-d20c-4233-99d3-98690412d5cd","Type":"ContainerStarted","Data":"62ce57d4f9e83125bc2b4ac8e884b9d1687b32ef0f39e8d772b19e48fe98339f"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.279122 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" event={"ID":"e04b8c46-159a-4536-8f53-faff85690626","Type":"ContainerStarted","Data":"5851399252dbff731837656afa1a2212de1fc87a43381b3bf7e20f5204c034a3"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.283309 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" event={"ID":"838ccf5b-b653-42f4-be39-b6cc8f451440","Type":"ContainerStarted","Data":"c4660343a0b6d7aec76bddf9a6ca7109d699f290998bf46ecc7d2797fd7e7afb"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.283367 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" event={"ID":"838ccf5b-b653-42f4-be39-b6cc8f451440","Type":"ContainerStarted","Data":"cf667ca2e97da4dba0a88771b214c3b7e674aaddb8d71047132e645b01525275"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.285185 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" event={"ID":"dfe9c513-2c23-4583-937f-4a44fcf0452a","Type":"ContainerStarted","Data":"1e45e36b0f62ed2a5c6cc5eb1ee77b2195fff53ab696b6148061cdb30f94528a"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.286781 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" event={"ID":"a4cf6b9e-f466-4d58-8ffd-85487c7a2531","Type":"ContainerStarted","Data":"8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.286805 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" event={"ID":"a4cf6b9e-f466-4d58-8ffd-85487c7a2531","Type":"ContainerStarted","Data":"ba1f211115ee5780bf36e05062363db4c2812fe062fea045039ea992dc12fe95"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.287347 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:42 crc kubenswrapper[4960]: W1002 07:18:42.289040 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69def292_5958_44c2_a8fb_362bbdb897a0.slice/crio-4fe6e8ffb21f1391ac33c465fc8d35ee586cce11cafb16f44d78655294cec265 WatchSource:0}: Error finding container 4fe6e8ffb21f1391ac33c465fc8d35ee586cce11cafb16f44d78655294cec265: Status 404 returned error can't find the container with id 4fe6e8ffb21f1391ac33c465fc8d35ee586cce11cafb16f44d78655294cec265 Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.299722 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" event={"ID":"2d62d400-9f39-4905-a5a2-bfbfe3753dd1","Type":"ContainerStarted","Data":"56fd05328cadfdd648b483efb2868cec67da8135be2d7d3b90997ad63f1ae818"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.309876 4960 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lddcs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.309948 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" podUID="a4cf6b9e-f466-4d58-8ffd-85487c7a2531" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.316085 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.379683 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.380866 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.880841837 +0000 UTC m=+143.912788124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.416570 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" event={"ID":"ec915743-c847-49f8-9d0e-c800b9fb2f2a","Type":"ContainerStarted","Data":"3512e182d0a49a4feb1262930c51fb6a1aedd92d1ea4f13f31b887c13d137763"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.416635 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-96r9w"] Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.416655 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nc4hj"] Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.416673 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" event={"ID":"ec915743-c847-49f8-9d0e-c800b9fb2f2a","Type":"ContainerStarted","Data":"728418e088af2e0f08533a51c1678279cc8ea84426437376a0fdbdfc0998c05e"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.483881 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.484084 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.984045468 +0000 UTC m=+144.015991755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.484312 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.486073 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:42.986062994 +0000 UTC m=+144.018009281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.496518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" event={"ID":"15752c4a-07e4-4fa6-9579-b5ac7ae26c12","Type":"ContainerStarted","Data":"a2417f970bc9441a6ded8d9baa5f3676adcb51373cbab19510e3463e5248fb53"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.508227 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" event={"ID":"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e","Type":"ContainerStarted","Data":"330065b54aa5549b7d66e3899366027c29c3b3cfcc029db5420f7940bba46ea8"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.508303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" event={"ID":"4e0d9ffd-9187-4446-9a5b-87b1b999ff6e","Type":"ContainerStarted","Data":"75d0f31e230408441e5f8b5289af91b7850307f7b24f7229cc9d4029ce492e70"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.516953 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" event={"ID":"ff8045d0-82c4-4054-8c26-f1bbf25dde9e","Type":"ContainerStarted","Data":"21161132df1bcf9e9f1067cf6203b32a639c56c8fd7da7cc3bc828226440263f"} Oct 02 07:18:42 crc kubenswrapper[4960]: W1002 07:18:42.533210 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97a0f863_045c_4251_865d_327d1c1397e5.slice/crio-16a8a614ddb1ed5dda660d698c8ec0b96b993c0ea3d7734839f4130a4dae7a45 WatchSource:0}: Error finding container 16a8a614ddb1ed5dda660d698c8ec0b96b993c0ea3d7734839f4130a4dae7a45: Status 404 returned error can't find the container with id 16a8a614ddb1ed5dda660d698c8ec0b96b993c0ea3d7734839f4130a4dae7a45 Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.583542 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" event={"ID":"be880348-4fad-4b89-8e00-6dd336636f76","Type":"ContainerStarted","Data":"38ad047e4cfcd97f201f66cbc3e153a827a013cad4e773e2c2044893f3cb7274"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.583608 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.586712 4960 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fr4q7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.586762 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" podUID="be880348-4fad-4b89-8e00-6dd336636f76" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.587852 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.589955 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.089926057 +0000 UTC m=+144.121872344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.611413 4960 generic.go:334] "Generic (PLEG): container finished" podID="0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6" containerID="091b01ebdeb267cf89c687b7963a0bf61d3e19c42b73ac6eee4c02680aa9358c" exitCode=0 Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.613379 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" event={"ID":"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6","Type":"ContainerDied","Data":"091b01ebdeb267cf89c687b7963a0bf61d3e19c42b73ac6eee4c02680aa9358c"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.620947 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" event={"ID":"e36d848d-519b-4e0f-bbc6-18fd6620ef2e","Type":"ContainerStarted","Data":"a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.621026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" event={"ID":"e36d848d-519b-4e0f-bbc6-18fd6620ef2e","Type":"ContainerStarted","Data":"bac05e546f09dd3315a11a2caaa608167331f17695309113be249122aa9f65ed"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.621618 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.649938 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rc25f" event={"ID":"7e2be296-6def-442a-8771-f9333e891d14","Type":"ContainerStarted","Data":"163f7dcbe503cc4e7f6cb18b3bdfebd1389c97200f61e326827ff163fedbbd9a"} Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.650263 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rc25f" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.674593 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-rc25f container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.674633 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rc25f" podUID="7e2be296-6def-442a-8771-f9333e891d14" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.692468 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.694057 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.194044268 +0000 UTC m=+144.225990555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.707877 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5"] Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.803377 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-89pc5" podStartSLOduration=121.803355468 podStartE2EDuration="2m1.803355468s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:42.801742955 +0000 UTC m=+143.833689242" watchObservedRunningTime="2025-10-02 07:18:42.803355468 +0000 UTC m=+143.835301755" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.804525 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.804869 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.304845996 +0000 UTC m=+144.336792283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.833203 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr"] Oct 02 07:18:42 crc kubenswrapper[4960]: W1002 07:18:42.894213 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1b94355_cfe1_447e_b2fe_dcc1730d7e39.slice/crio-945d8e218eb056fd7167c831637fa7db695919f78bc6bc6b21352522cb9a12a5 WatchSource:0}: Error finding container 945d8e218eb056fd7167c831637fa7db695919f78bc6bc6b21352522cb9a12a5: Status 404 returned error can't find the container with id 945d8e218eb056fd7167c831637fa7db695919f78bc6bc6b21352522cb9a12a5 Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.912814 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.913143 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.916330 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" podStartSLOduration=121.916271364 podStartE2EDuration="2m1.916271364s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:42.912161531 +0000 UTC m=+143.944107838" watchObservedRunningTime="2025-10-02 07:18:42.916271364 +0000 UTC m=+143.948217681" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.917226 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:42 crc kubenswrapper[4960]: E1002 07:18:42.918104 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.418079292 +0000 UTC m=+144.450025769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:42 crc kubenswrapper[4960]: W1002 07:18:42.926383 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57150513_840e_47ef_8e38_91b3b090a380.slice/crio-a04881b94b5145f28f44b2e9ca06596abc0bc03eb4fb150a47ff645ecc4bdf33 WatchSource:0}: Error finding container a04881b94b5145f28f44b2e9ca06596abc0bc03eb4fb150a47ff645ecc4bdf33: Status 404 returned error can't find the container with id a04881b94b5145f28f44b2e9ca06596abc0bc03eb4fb150a47ff645ecc4bdf33 Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.955616 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q2x4f" podStartSLOduration=121.955587677 podStartE2EDuration="2m1.955587677s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:42.952807487 +0000 UTC m=+143.984753774" watchObservedRunningTime="2025-10-02 07:18:42.955587677 +0000 UTC m=+143.987533964" Oct 02 07:18:42 crc kubenswrapper[4960]: I1002 07:18:42.986039 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd"] Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.022433 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z"] Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.026387 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.026951 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.526922007 +0000 UTC m=+144.558868294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.101729 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dgl58"] Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.131490 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.143242 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.643220373 +0000 UTC m=+144.675166660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.232641 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.233025 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.73300532 +0000 UTC m=+144.764951607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.334661 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.335034 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.835020783 +0000 UTC m=+144.866967080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.448769 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.449175 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:43.949142279 +0000 UTC m=+144.981088566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.502572 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7r26m" podStartSLOduration=122.502552748 podStartE2EDuration="2m2.502552748s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:43.501492204 +0000 UTC m=+144.533438491" watchObservedRunningTime="2025-10-02 07:18:43.502552748 +0000 UTC m=+144.534499035" Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.551298 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.551757 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.051739991 +0000 UTC m=+145.083686278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.654774 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.655137 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.155115939 +0000 UTC m=+145.187062226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.702572 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-52fm9" podStartSLOduration=122.702554305 podStartE2EDuration="2m2.702554305s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:43.701652895 +0000 UTC m=+144.733599182" watchObservedRunningTime="2025-10-02 07:18:43.702554305 +0000 UTC m=+144.734500592" Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.702988 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-jzg6w" podStartSLOduration=122.702968138 podStartE2EDuration="2m2.702968138s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:43.647162941 +0000 UTC m=+144.679109228" watchObservedRunningTime="2025-10-02 07:18:43.702968138 +0000 UTC m=+144.734914425" Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.743330 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" podStartSLOduration=122.743311074 podStartE2EDuration="2m2.743311074s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:43.741573957 +0000 UTC m=+144.773520244" watchObservedRunningTime="2025-10-02 07:18:43.743311074 +0000 UTC m=+144.775257361" Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.756775 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.757220 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.257205664 +0000 UTC m=+145.289151951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.784621 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" event={"ID":"be880348-4fad-4b89-8e00-6dd336636f76","Type":"ContainerStarted","Data":"3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0"} Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.853961 4960 generic.go:334] "Generic (PLEG): container finished" podID="ff8045d0-82c4-4054-8c26-f1bbf25dde9e" containerID="482462b203487ef93f60d7292916c4b45d30330a7004b990d384df6d14702dca" exitCode=0 Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.854122 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" event={"ID":"ff8045d0-82c4-4054-8c26-f1bbf25dde9e","Type":"ContainerDied","Data":"482462b203487ef93f60d7292916c4b45d30330a7004b990d384df6d14702dca"} Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.859644 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.860079 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.360058734 +0000 UTC m=+145.392005021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.872517 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" event={"ID":"be65bb8e-a260-43f7-8e00-78c4158a683e","Type":"ContainerStarted","Data":"215ced1dcf7f51e6080b60f01a598fba6cd5e17e2c69e5d8ed9ec541ff3dc7ed"} Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.924425 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" event={"ID":"69def292-5958-44c2-a8fb-362bbdb897a0","Type":"ContainerStarted","Data":"e18873f1fe3e26c0a8d9647c4cc4442e03727fd9c965b88f1afa97ee9ff00876"} Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.924500 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" event={"ID":"69def292-5958-44c2-a8fb-362bbdb897a0","Type":"ContainerStarted","Data":"4fe6e8ffb21f1391ac33c465fc8d35ee586cce11cafb16f44d78655294cec265"} Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.938086 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bpfrh" event={"ID":"f1b94355-cfe1-447e-b2fe-dcc1730d7e39","Type":"ContainerStarted","Data":"945d8e218eb056fd7167c831637fa7db695919f78bc6bc6b21352522cb9a12a5"} Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.944769 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.947904 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zcpbw" podStartSLOduration=121.947873248 podStartE2EDuration="2m1.947873248s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:43.945603064 +0000 UTC m=+144.977549351" watchObservedRunningTime="2025-10-02 07:18:43.947873248 +0000 UTC m=+144.979819535" Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.958230 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vz5gr" event={"ID":"a5911cf0-70ba-44a3-8c49-4cd66cca73b7","Type":"ContainerStarted","Data":"5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22"} Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.961641 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:43 crc kubenswrapper[4960]: E1002 07:18:43.964673 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.464655542 +0000 UTC m=+145.496601829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.981855 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" podStartSLOduration=121.981830577 podStartE2EDuration="2m1.981830577s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:43.980444233 +0000 UTC m=+145.012390520" watchObservedRunningTime="2025-10-02 07:18:43.981830577 +0000 UTC m=+145.013776864" Oct 02 07:18:43 crc kubenswrapper[4960]: I1002 07:18:43.998179 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" event={"ID":"b130afa1-783b-4b51-98bb-567c0d92830e","Type":"ContainerStarted","Data":"16047fc9026cf4c301e09944d6f43d54c9f2113f4b6b53ae091a0893ef158f0c"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.018968 4960 generic.go:334] "Generic (PLEG): container finished" podID="2d62d400-9f39-4905-a5a2-bfbfe3753dd1" containerID="e7915087e4b01aa0a2681a39bb03d3c466ebfae3b5106c91331322dd2fe01cde" exitCode=0 Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.019126 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" event={"ID":"2d62d400-9f39-4905-a5a2-bfbfe3753dd1","Type":"ContainerDied","Data":"e7915087e4b01aa0a2681a39bb03d3c466ebfae3b5106c91331322dd2fe01cde"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.028839 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" event={"ID":"d1070796-a85a-4cd5-a302-bdaa2541db68","Type":"ContainerStarted","Data":"72ee8211e75c17fd9381e5b8d97c31d8eddfcd27154083f5858ab5c009539004"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.030166 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" event={"ID":"0122bd5f-ff81-4468-aac9-32f277f12e68","Type":"ContainerStarted","Data":"59a97a7f46fb0f0e705e6caee2ef204e07004914c23cc4722493d121b465c9e9"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.030198 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" event={"ID":"0122bd5f-ff81-4468-aac9-32f277f12e68","Type":"ContainerStarted","Data":"a59dfca4fb7b9d21d63f71b2f420e14eeb8966a957cc7cae1e1b31bbac21feed"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.031525 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.042775 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" event={"ID":"c3480a6e-4787-4d30-b05b-da16f1356be4","Type":"ContainerStarted","Data":"718f55556ec3d12407651d669eb1e6200c3b8c89cc465c4ececee3f31264e53d"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.056325 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" event={"ID":"3cda2a13-5fea-4023-afb5-ed377f21b98e","Type":"ContainerStarted","Data":"45eed7117be90a7647e4d07406015783c91809db970e83418cd408f5e8416d6a"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.061279 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" event={"ID":"e34b1da7-5b94-4313-81f8-401e1a798a09","Type":"ContainerStarted","Data":"f685fe3ccf00fde97653dcd63c3226e23053a794fe874457c742e6d892d687e1"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.062302 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.064246 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" event={"ID":"57150513-840e-47ef-8e38-91b3b090a380","Type":"ContainerStarted","Data":"a04881b94b5145f28f44b2e9ca06596abc0bc03eb4fb150a47ff645ecc4bdf33"} Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.064478 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.564454623 +0000 UTC m=+145.596400910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.079173 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-rc25f" podStartSLOduration=123.079147179 podStartE2EDuration="2m3.079147179s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.074925792 +0000 UTC m=+145.106872089" watchObservedRunningTime="2025-10-02 07:18:44.079147179 +0000 UTC m=+145.111093466" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.084808 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" event={"ID":"97a0f863-045c-4251-865d-327d1c1397e5","Type":"ContainerStarted","Data":"16a8a614ddb1ed5dda660d698c8ec0b96b993c0ea3d7734839f4130a4dae7a45"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.094149 4960 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wdkwq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:5443/healthz\": dial tcp 10.217.0.18:5443: connect: connection refused" start-of-body= Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.094229 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" podUID="69def292-5958-44c2-a8fb-362bbdb897a0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.18:5443/healthz\": dial tcp 10.217.0.18:5443: connect: connection refused" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.095638 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" event={"ID":"e04b8c46-159a-4536-8f53-faff85690626","Type":"ContainerStarted","Data":"cd5041b1f38908cceeb153355e9e5d1d35599be52a73592069535f564b5a1484"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.106401 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" event={"ID":"ec915743-c847-49f8-9d0e-c800b9fb2f2a","Type":"ContainerStarted","Data":"4e776425ca1131a0f04cf4f444ec8b2de07bd519b14c372594516b43f07fb718"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.107151 4960 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-96r9w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.107188 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" podUID="0122bd5f-ff81-4468-aac9-32f277f12e68" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.113756 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9lf9c" event={"ID":"f6b93888-4d26-4500-ae07-aefff1dc0a7b","Type":"ContainerStarted","Data":"203f51b78519be0f30b9fa0e80b570c6f841077ac20a33def0d2401583b5b61a"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.113815 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9lf9c" event={"ID":"f6b93888-4d26-4500-ae07-aefff1dc0a7b","Type":"ContainerStarted","Data":"82ff591f6cc330bc4c74347d5a71ab38a70821980793c12ddf32a999c1981260"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.131138 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" podStartSLOduration=122.131111572 podStartE2EDuration="2m2.131111572s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.116768776 +0000 UTC m=+145.148715063" watchObservedRunningTime="2025-10-02 07:18:44.131111572 +0000 UTC m=+145.163057849" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.131532 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.144582 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" event={"ID":"dfe9c513-2c23-4583-937f-4a44fcf0452a","Type":"ContainerStarted","Data":"e747b5c097bb712b967fbb8d087d7dd79db0e24c2196f7534951a827a902c8ba"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.161353 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" event={"ID":"c019958e-ed31-4dc5-b8f7-135cf1e69f70","Type":"ContainerStarted","Data":"d41cd4a6c725f5bbcccd86dd1847cc39793086c1c58b3a5384710119ce168155"} Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.161997 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-rc25f container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.162065 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rc25f" podUID="7e2be296-6def-442a-8771-f9333e891d14" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.164253 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.165391 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-smfhg" podStartSLOduration=123.165375311 podStartE2EDuration="2m3.165375311s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.162854489 +0000 UTC m=+145.194800776" watchObservedRunningTime="2025-10-02 07:18:44.165375311 +0000 UTC m=+145.197321598" Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.175892 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.675871011 +0000 UTC m=+145.707817298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.226460 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" podStartSLOduration=123.226437068 podStartE2EDuration="2m3.226437068s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.212497997 +0000 UTC m=+145.244444284" watchObservedRunningTime="2025-10-02 07:18:44.226437068 +0000 UTC m=+145.258383355" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.249138 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.265189 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.265397 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.765335188 +0000 UTC m=+145.797281485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.265933 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.275186 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.775143496 +0000 UTC m=+145.807090003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.275861 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" podStartSLOduration=122.275831737 podStartE2EDuration="2m2.275831737s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.264297954 +0000 UTC m=+145.296244241" watchObservedRunningTime="2025-10-02 07:18:44.275831737 +0000 UTC m=+145.307778024" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.291200 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.309643 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.333879 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xxnvh" podStartSLOduration=123.333847586 podStartE2EDuration="2m3.333847586s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.324626468 +0000 UTC m=+145.356572785" watchObservedRunningTime="2025-10-02 07:18:44.333847586 +0000 UTC m=+145.365793873" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.384029 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.384909 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.884868768 +0000 UTC m=+145.916815055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.385143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.385572 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.88556271 +0000 UTC m=+145.917508997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.393253 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9lf9c" podStartSLOduration=123.393221759 podStartE2EDuration="2m3.393221759s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.361566443 +0000 UTC m=+145.393512730" watchObservedRunningTime="2025-10-02 07:18:44.393221759 +0000 UTC m=+145.425168066" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.417636 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.418057 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.418156 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.488996 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.489377 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:44.989354912 +0000 UTC m=+146.021301209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.497474 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7sh7k" podStartSLOduration=123.497445373 podStartE2EDuration="2m3.497445373s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.48868946 +0000 UTC m=+145.520635747" watchObservedRunningTime="2025-10-02 07:18:44.497445373 +0000 UTC m=+145.529391660" Oct 02 07:18:44 crc kubenswrapper[4960]: W1002 07:18:44.540750 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5926160e_8baa_4e5b_b0ce_2de658e7250f.slice/crio-098fd9a179df5a04eaf28afde409c5504e5dff81214cc29d73fbda77f1045586 WatchSource:0}: Error finding container 098fd9a179df5a04eaf28afde409c5504e5dff81214cc29d73fbda77f1045586: Status 404 returned error can't find the container with id 098fd9a179df5a04eaf28afde409c5504e5dff81214cc29d73fbda77f1045586 Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.592628 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.593582 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.093567016 +0000 UTC m=+146.125513293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.621762 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vz5gr" podStartSLOduration=123.621729978 podStartE2EDuration="2m3.621729978s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.586816627 +0000 UTC m=+145.618762914" watchObservedRunningTime="2025-10-02 07:18:44.621729978 +0000 UTC m=+145.653676265" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.689639 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nnbxf"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.690225 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9vwr2" podStartSLOduration=123.690200075 podStartE2EDuration="2m3.690200075s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:44.642655125 +0000 UTC m=+145.674601412" watchObservedRunningTime="2025-10-02 07:18:44.690200075 +0000 UTC m=+145.722146362" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.694573 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.694906 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.194886236 +0000 UTC m=+146.226832523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.757585 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.796792 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.797155 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.297142757 +0000 UTC m=+146.329089034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.901091 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:44 crc kubenswrapper[4960]: E1002 07:18:44.904165 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.404131572 +0000 UTC m=+146.436077859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.956739 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.960165 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.967252 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:44 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:44 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:44 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.967743 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.976200 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gzvxm"] Oct 02 07:18:44 crc kubenswrapper[4960]: I1002 07:18:44.986520 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gpq9m"] Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.008059 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.008409 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.508396358 +0000 UTC m=+146.540342645 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.033675 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xkfg8"] Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.116895 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.117166 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.61714646 +0000 UTC m=+146.649092747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.117270 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.117683 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.617661256 +0000 UTC m=+146.649607543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.220444 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.220814 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.720794586 +0000 UTC m=+146.752740873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.249900 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gpq9m" event={"ID":"f58114fb-9f32-45ab-9a99-5c4a378353de","Type":"ContainerStarted","Data":"61380b4b7d4f730ad22b798621ce32a3bf7ef3d449fc6f7555d86be3a00a9530"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.255425 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" event={"ID":"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34","Type":"ContainerStarted","Data":"8629f65b90a45965997d0b61036d036f379fef58dcb0a60595daf7b2ef33cac6"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.264900 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" event={"ID":"5926160e-8baa-4e5b-b0ce-2de658e7250f","Type":"ContainerStarted","Data":"098fd9a179df5a04eaf28afde409c5504e5dff81214cc29d73fbda77f1045586"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.311943 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" event={"ID":"be65bb8e-a260-43f7-8e00-78c4158a683e","Type":"ContainerStarted","Data":"1bd73e5a446a656d1faf596f517a836448363b7387c7758a9088b7511949ecce"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.312799 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.324811 4960 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-2h9xd container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.324880 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" podUID="be65bb8e-a260-43f7-8e00-78c4158a683e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.326370 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.326664 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.826652594 +0000 UTC m=+146.858598881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.340572 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" podStartSLOduration=123.340556053 podStartE2EDuration="2m3.340556053s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.337595807 +0000 UTC m=+146.369542094" watchObservedRunningTime="2025-10-02 07:18:45.340556053 +0000 UTC m=+146.372502340" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.353928 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" event={"ID":"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6","Type":"ContainerStarted","Data":"6bec8f66b3d9037dcfe8b8b59190ba28be9611d73c2e4b2f8c354588b1aec7f1"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.370328 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" event={"ID":"dffe37b6-aae9-49ff-a8a7-2c64360d748b","Type":"ContainerStarted","Data":"82213b675b1d46f4b3d6dc0eb7295f2aa99d707ebdd652b38027c1518c08006e"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.413165 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" event={"ID":"57150513-840e-47ef-8e38-91b3b090a380","Type":"ContainerStarted","Data":"c4efd57b1395c4d59223f46957df11ea6ecceba58e22a79b51835ea96daf948f"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.423510 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" event={"ID":"ff8045d0-82c4-4054-8c26-f1bbf25dde9e","Type":"ContainerStarted","Data":"2d0d6267fd653edc61447a9f8ee720da949860761ba2034a9594430b760e6735"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.423610 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.426252 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" event={"ID":"08abc337-7b21-4bc4-8e31-7b594d654312","Type":"ContainerStarted","Data":"6df73252e31312b30356fa2f085030e10e358cb73aa67721b5a7a0157acf0f8d"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.427604 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.427828 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.927800359 +0000 UTC m=+146.959746646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.428497 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.429219 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:45.929207414 +0000 UTC m=+146.961153701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.434430 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" event={"ID":"c3480a6e-4787-4d30-b05b-da16f1356be4","Type":"ContainerStarted","Data":"da65758b8f7139dae33b959d5bd6820ebb936451f6effecab47cca864bae32fd"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.435434 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.443707 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" event={"ID":"97a0f863-045c-4251-865d-327d1c1397e5","Type":"ContainerStarted","Data":"be069d06de2e3d9000a4f6cd982946be60106145e944ec354eed034347a8211c"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.448657 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.454234 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" podStartSLOduration=124.454220915 podStartE2EDuration="2m4.454220915s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.453245843 +0000 UTC m=+146.485192130" watchObservedRunningTime="2025-10-02 07:18:45.454220915 +0000 UTC m=+146.486167202" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.458665 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" event={"ID":"c019958e-ed31-4dc5-b8f7-135cf1e69f70","Type":"ContainerStarted","Data":"7fcc4f39c19f61b3272037dc1b80c79593b9edd427b99a54d092c5ed921f30c4"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.464726 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xkfg8" event={"ID":"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932","Type":"ContainerStarted","Data":"162a04ec3d54a268b9e40d3064039562e23ae81aaea68189cf2e09adc11a73c6"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.493357 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qcf8z" podStartSLOduration=123.493336041 podStartE2EDuration="2m3.493336041s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.48281347 +0000 UTC m=+146.514759757" watchObservedRunningTime="2025-10-02 07:18:45.493336041 +0000 UTC m=+146.525282328" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.530724 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.532216 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.032190669 +0000 UTC m=+147.064136956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.561449 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dgl58" podStartSLOduration=123.561419625 podStartE2EDuration="2m3.561419625s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.539204186 +0000 UTC m=+146.571150483" watchObservedRunningTime="2025-10-02 07:18:45.561419625 +0000 UTC m=+146.593365932" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.573065 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" event={"ID":"3cda2a13-5fea-4023-afb5-ed377f21b98e","Type":"ContainerStarted","Data":"d1186ea4621c9cfdef4b974daba95b889a6bfef0bc38b99022d8d046c46c1e9a"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.640760 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.641699 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.141678465 +0000 UTC m=+147.173624752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.688006 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bpfrh" event={"ID":"f1b94355-cfe1-447e-b2fe-dcc1730d7e39","Type":"ContainerStarted","Data":"297c2f18b128d455083446146283cdb51697ca60e24257f708166ce313623bce"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.689999 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-nc4hj" podStartSLOduration=123.689989428 podStartE2EDuration="2m3.689989428s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.639613147 +0000 UTC m=+146.671559434" watchObservedRunningTime="2025-10-02 07:18:45.689989428 +0000 UTC m=+146.721935715" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.735458 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" event={"ID":"3d80dacd-d5ee-4616-878d-0563d69b6b6f","Type":"ContainerStarted","Data":"30bb1ea52d3ee5fd6226cc94c7fe95f5ad8388144c755fcec61e6a4c552d71dc"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.740322 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" event={"ID":"e919b925-67b0-475f-b01c-e37ae52c3dbb","Type":"ContainerStarted","Data":"56b267e03976a07338645a43603c11644f0142175fbf7d80815ace111269af88"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.741903 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.743450 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.242317342 +0000 UTC m=+147.274263629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.743603 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.743893 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.243883584 +0000 UTC m=+147.275829871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.753165 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" event={"ID":"d1070796-a85a-4cd5-a302-bdaa2541db68","Type":"ContainerStarted","Data":"fd4c32d8d25cdb079a1a7474d809dd0c3841abf11b3673fdd0fa7ee896e7a0fd"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.756440 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bpfrh" podStartSLOduration=6.756423649 podStartE2EDuration="6.756423649s" podCreationTimestamp="2025-10-02 07:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.756115579 +0000 UTC m=+146.788061866" watchObservedRunningTime="2025-10-02 07:18:45.756423649 +0000 UTC m=+146.788369936" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.796584 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-h6vf5" podStartSLOduration=124.796566259 podStartE2EDuration="2m4.796566259s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.795337269 +0000 UTC m=+146.827283556" watchObservedRunningTime="2025-10-02 07:18:45.796566259 +0000 UTC m=+146.828512546" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.812761 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" event={"ID":"e34b1da7-5b94-4313-81f8-401e1a798a09","Type":"ContainerStarted","Data":"0d288b20e52939b78236cf8392ad30575ad3975c3775c9badbb7950ac745817f"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.817389 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" podStartSLOduration=123.817370533 podStartE2EDuration="2m3.817370533s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.816440693 +0000 UTC m=+146.848386980" watchObservedRunningTime="2025-10-02 07:18:45.817370533 +0000 UTC m=+146.849316820" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.844646 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.846904 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.346868498 +0000 UTC m=+147.378814785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.852521 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z8gjv" podStartSLOduration=124.852504741 podStartE2EDuration="2m4.852504741s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.850892849 +0000 UTC m=+146.882839136" watchObservedRunningTime="2025-10-02 07:18:45.852504741 +0000 UTC m=+146.884451028" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.876936 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" event={"ID":"b4ffa2a8-76aa-48e8-a617-597c63719217","Type":"ContainerStarted","Data":"5543d995610c731dee1c2b8edc420c9de90e94c6da89544913467f94d50b4e09"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.877617 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" event={"ID":"b4ffa2a8-76aa-48e8-a617-597c63719217","Type":"ContainerStarted","Data":"376bb2d349401fc05b374a851d8288c551862da42d4d32980bd03fdfeafad8ed"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.910180 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" event={"ID":"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327","Type":"ContainerStarted","Data":"dfba0c44d4e4687b1f60f49bdcd6534925721394d11bee3fc38e6310bec69f96"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.941902 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" event={"ID":"fef811a3-779c-4bf7-8eb2-cbab93f43f8e","Type":"ContainerStarted","Data":"5473f869bd0d8992c2de219afd2fe6a390eedb93ce8eb4b7db1a1357d3bbd5a2"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.957143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:45 crc kubenswrapper[4960]: E1002 07:18:45.957441 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.457426508 +0000 UTC m=+147.489372795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.959874 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" event={"ID":"2d62d400-9f39-4905-a5a2-bfbfe3753dd1","Type":"ContainerStarted","Data":"089a7ac235507f281b4b5a65e14a7c6b55f2982fde086772d626c1677883f753"} Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.964768 4960 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-96r9w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.964817 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" podUID="0122bd5f-ff81-4468-aac9-32f277f12e68" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.978447 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:45 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:45 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:45 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.978534 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:45 crc kubenswrapper[4960]: I1002 07:18:45.993912 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdkwq" Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:45.999605 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" podStartSLOduration=123.999586423 podStartE2EDuration="2m3.999586423s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:45.998198258 +0000 UTC m=+147.030144545" watchObservedRunningTime="2025-10-02 07:18:45.999586423 +0000 UTC m=+147.031532710" Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.020554 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.020610 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.061436 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.062427 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.562395117 +0000 UTC m=+147.594341404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.190909 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.191733 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.691710084 +0000 UTC m=+147.723656431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.292898 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.293705 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.793681066 +0000 UTC m=+147.825627353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.394637 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.395041 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.895028028 +0000 UTC m=+147.926974315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.495439 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.495868 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:46.995846493 +0000 UTC m=+148.027792780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.596715 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.597348 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.097135302 +0000 UTC m=+148.129081589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.698261 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.698663 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.198642649 +0000 UTC m=+148.230588936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.799673 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.800119 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.300101404 +0000 UTC m=+148.332047691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.901156 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:46 crc kubenswrapper[4960]: E1002 07:18:46.901496 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.401476467 +0000 UTC m=+148.433422754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.964665 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:46 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:46 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:46 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.965058 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.966129 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4xjvf" event={"ID":"3d80dacd-d5ee-4616-878d-0563d69b6b6f","Type":"ContainerStarted","Data":"ba9f53650d8a285bfe0ac444f32864886750cf193019da0afbb83f5ad6e49629"} Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.967678 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" event={"ID":"fef811a3-779c-4bf7-8eb2-cbab93f43f8e","Type":"ContainerStarted","Data":"ac4f653bf393c12c0c2f1491fa184d4a8aba9bd724436b1ff9b8325bee1f35b8"} Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.969580 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" event={"ID":"0b2ae377-b5f0-4c98-85f6-fa1923b7a7c6","Type":"ContainerStarted","Data":"72d68b63fcd7e2e23eef5deff6b2b24526f754eb58344b043f15e3fbe907f088"} Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.971221 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" event={"ID":"e919b925-67b0-475f-b01c-e37ae52c3dbb","Type":"ContainerStarted","Data":"4fe4759c04157fce381000b23d7ef2f6d5af1598af8c2ce729a4fa488a01afc0"} Oct 02 07:18:46 crc kubenswrapper[4960]: I1002 07:18:46.982823 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" event={"ID":"57150513-840e-47ef-8e38-91b3b090a380","Type":"ContainerStarted","Data":"b2f7ddeb42a8ed49f4ab81b1024a210872b2bed71c4781212f9511822a0d3f89"} Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.004406 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.504387259 +0000 UTC m=+148.536333546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.003870 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.008373 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" event={"ID":"08abc337-7b21-4bc4-8e31-7b594d654312","Type":"ContainerStarted","Data":"344f63c698d3586035b3131e78bd29ad2d8ef930da505b3fb62803251b0e853c"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.008424 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" event={"ID":"08abc337-7b21-4bc4-8e31-7b594d654312","Type":"ContainerStarted","Data":"d26f695bb7bcc608e00d7f302e5818d533d8c66488760ed590072bf1ccc3b52a"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.020473 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" event={"ID":"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34","Type":"ContainerStarted","Data":"2d15e7272e398e8b3fb423c09119f57ba2d9d48e1f2b4736fdfb01398687f57d"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.020529 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" event={"ID":"7aa2fa9d-f38b-4562-bf5d-cbcfdafb2d34","Type":"ContainerStarted","Data":"d3fc2e14ec27ee764e500649039bb9d8407e51185e6fcf1bb7b158e745147578"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.021219 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.033086 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" event={"ID":"dffe37b6-aae9-49ff-a8a7-2c64360d748b","Type":"ContainerStarted","Data":"d833812d7afa07661166e668b5fe3a21830ab88847df5bfcd449f8e93be969e8"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.033145 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" event={"ID":"dffe37b6-aae9-49ff-a8a7-2c64360d748b","Type":"ContainerStarted","Data":"0b34046a5d29c91b6664c0cd2de466f50ce8e7c9fcbaea509d58ea6a331db07c"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.042634 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gpq9m" event={"ID":"f58114fb-9f32-45ab-9a99-5c4a378353de","Type":"ContainerStarted","Data":"4e214ff10d1b3e785673b95a3d24b8db5d2c8a53fab48e9b5a7d0156080bc189"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.050988 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zbqtp" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.051227 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xkfg8" event={"ID":"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932","Type":"ContainerStarted","Data":"0d50a965174c952405babf2b587c7ff692cc64a738e0561aad9d0a2f5153c99f"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.056613 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" event={"ID":"b4ffa2a8-76aa-48e8-a617-597c63719217","Type":"ContainerStarted","Data":"3a62aeacb45fe842633f33e50d4b49711d5891b25878b0343cdbcfcebe8aa332"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.059302 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" podStartSLOduration=126.059290137 podStartE2EDuration="2m6.059290137s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.055241096 +0000 UTC m=+148.087187403" watchObservedRunningTime="2025-10-02 07:18:47.059290137 +0000 UTC m=+148.091236424" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.061247 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" event={"ID":"5926160e-8baa-4e5b-b0ce-2de658e7250f","Type":"ContainerStarted","Data":"6c01bbec3b794159b6b257fec555534da9e5baa3114012d48c41a22a92406a42"} Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.080303 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2h9xd" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.103512 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xkvwr" podStartSLOduration=125.103488928 podStartE2EDuration="2m5.103488928s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.102834077 +0000 UTC m=+148.134780364" watchObservedRunningTime="2025-10-02 07:18:47.103488928 +0000 UTC m=+148.135435215" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.109546 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.111462 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.611438975 +0000 UTC m=+148.643385262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.212328 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.223920 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmx7" podStartSLOduration=125.223902987 podStartE2EDuration="2m5.223902987s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.193295746 +0000 UTC m=+148.225242033" watchObservedRunningTime="2025-10-02 07:18:47.223902987 +0000 UTC m=+148.255849274" Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.233008 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.732966231 +0000 UTC m=+148.764912528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.269223 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mrbf" podStartSLOduration=125.269204154 podStartE2EDuration="2m5.269204154s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.268666637 +0000 UTC m=+148.300612924" watchObservedRunningTime="2025-10-02 07:18:47.269204154 +0000 UTC m=+148.301150441" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.305128 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xwqkr"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.306073 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.314373 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.314565 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.314928 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.814911464 +0000 UTC m=+148.846857751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.317167 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9lz9x" podStartSLOduration=125.317142086 podStartE2EDuration="2m5.317142086s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.306186781 +0000 UTC m=+148.338133068" watchObservedRunningTime="2025-10-02 07:18:47.317142086 +0000 UTC m=+148.349088373" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.321197 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xwqkr"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.360173 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-gpq9m" podStartSLOduration=9.360151879 podStartE2EDuration="9.360151879s" podCreationTimestamp="2025-10-02 07:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.359626151 +0000 UTC m=+148.391572448" watchObservedRunningTime="2025-10-02 07:18:47.360151879 +0000 UTC m=+148.392098166" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.415682 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-utilities\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.415752 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58pvm\" (UniqueName: \"kubernetes.io/projected/92ef421d-9909-46b7-a040-4fecbc24ec8e-kube-api-access-58pvm\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.415812 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.415835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-catalog-content\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.416220 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:47.916207754 +0000 UTC m=+148.948154041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.428560 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" podStartSLOduration=125.428540474 podStartE2EDuration="2m5.428540474s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.426807177 +0000 UTC m=+148.458753464" watchObservedRunningTime="2025-10-02 07:18:47.428540474 +0000 UTC m=+148.460486761" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.484786 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ww8hs"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.485923 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.492768 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.515527 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ww8hs"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.517486 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.517798 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-utilities\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.517839 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.517864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58pvm\" (UniqueName: \"kubernetes.io/projected/92ef421d-9909-46b7-a040-4fecbc24ec8e-kube-api-access-58pvm\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.517891 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.517913 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.517957 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-catalog-content\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.518362 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-catalog-content\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.519897 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.520220 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.02017524 +0000 UTC m=+149.052121647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.520516 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-utilities\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.534947 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.539401 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.580920 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.596923 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58pvm\" (UniqueName: \"kubernetes.io/projected/92ef421d-9909-46b7-a040-4fecbc24ec8e-kube-api-access-58pvm\") pod \"community-operators-xwqkr\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.600417 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9r8xz" podStartSLOduration=126.600398118 podStartE2EDuration="2m6.600398118s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.599748347 +0000 UTC m=+148.631694644" watchObservedRunningTime="2025-10-02 07:18:47.600398118 +0000 UTC m=+148.632344415" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.636170 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.641481 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-catalog-content\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.641594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.641648 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-utilities\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.641781 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prb5t\" (UniqueName: \"kubernetes.io/projected/c816bb24-42ce-4155-906b-3817beeb1693-kube-api-access-prb5t\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.641817 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.642634 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.142620506 +0000 UTC m=+149.174566793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.691294 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sznw8" podStartSLOduration=126.69127454 podStartE2EDuration="2m6.69127454s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:47.642534653 +0000 UTC m=+148.674480940" watchObservedRunningTime="2025-10-02 07:18:47.69127454 +0000 UTC m=+148.723220827" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.691931 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.698711 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xvrdl"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.699893 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.729338 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvrdl"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.745941 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.746149 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-catalog-content\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.746188 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-utilities\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.746225 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prb5t\" (UniqueName: \"kubernetes.io/projected/c816bb24-42ce-4155-906b-3817beeb1693-kube-api-access-prb5t\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.746703 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.246682775 +0000 UTC m=+149.278629052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.747560 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-catalog-content\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.747672 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-utilities\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.754875 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.763491 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.780459 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prb5t\" (UniqueName: \"kubernetes.io/projected/c816bb24-42ce-4155-906b-3817beeb1693-kube-api-access-prb5t\") pod \"certified-operators-ww8hs\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.803820 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.849433 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.849491 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97sc7\" (UniqueName: \"kubernetes.io/projected/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-kube-api-access-97sc7\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.849553 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-utilities\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.849580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-catalog-content\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.850107 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.350069703 +0000 UTC m=+149.382015990 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.850542 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.889536 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fqsbz"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.891052 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.932443 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqsbz"] Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.950896 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.951189 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-utilities\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.951263 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-catalog-content\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.951342 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97sc7\" (UniqueName: \"kubernetes.io/projected/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-kube-api-access-97sc7\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: E1002 07:18:47.951880 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.451858438 +0000 UTC m=+149.483804725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.952462 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-utilities\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.952788 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-catalog-content\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.970991 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:47 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:47 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:47 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.971069 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:47 crc kubenswrapper[4960]: I1002 07:18:47.978680 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97sc7\" (UniqueName: \"kubernetes.io/projected/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-kube-api-access-97sc7\") pod \"community-operators-xvrdl\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.053024 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-catalog-content\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.053473 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.053500 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-utilities\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.053591 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2gzl\" (UniqueName: \"kubernetes.io/projected/f8438dbe-9de4-42cc-ae67-fc5368e55f58-kube-api-access-d2gzl\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.053880 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.553867241 +0000 UTC m=+149.585813528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.063863 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.114605 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" event={"ID":"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327","Type":"ContainerStarted","Data":"bf52f7c5a46b44bde59dd4fa24282d7b7dc2e127930d79ae095a7428e319fa13"} Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.153868 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" event={"ID":"fef811a3-779c-4bf7-8eb2-cbab93f43f8e","Type":"ContainerStarted","Data":"3a1ce79e61d42515057aea23f4a3b6d66e81a3481f5821d4d439e31c9e08ef86"} Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.154474 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.154832 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2gzl\" (UniqueName: \"kubernetes.io/projected/f8438dbe-9de4-42cc-ae67-fc5368e55f58-kube-api-access-d2gzl\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.154864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-catalog-content\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.154897 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-utilities\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.155340 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-utilities\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.155424 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.6554052 +0000 UTC m=+149.687351487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.156036 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-catalog-content\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.187995 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xkfg8" event={"ID":"27d5c5ed-4a15-4889-9d0f-8fca8ba4f932","Type":"ContainerStarted","Data":"c594ae31a4c0bc1d212c74b20d931fc86e997adfab31bbe2e5d44bb97b68b4db"} Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.188736 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.218678 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-gzvxm" podStartSLOduration=126.218659028 podStartE2EDuration="2m6.218659028s" podCreationTimestamp="2025-10-02 07:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:48.207437854 +0000 UTC m=+149.239384151" watchObservedRunningTime="2025-10-02 07:18:48.218659028 +0000 UTC m=+149.250605315" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.226473 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2gzl\" (UniqueName: \"kubernetes.io/projected/f8438dbe-9de4-42cc-ae67-fc5368e55f58-kube-api-access-d2gzl\") pod \"certified-operators-fqsbz\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.235898 4960 generic.go:334] "Generic (PLEG): container finished" podID="97a0f863-045c-4251-865d-327d1c1397e5" containerID="be069d06de2e3d9000a4f6cd982946be60106145e944ec354eed034347a8211c" exitCode=0 Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.236804 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" event={"ID":"97a0f863-045c-4251-865d-327d1c1397e5","Type":"ContainerDied","Data":"be069d06de2e3d9000a4f6cd982946be60106145e944ec354eed034347a8211c"} Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.261074 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.262887 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.7628679 +0000 UTC m=+149.794814187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.263409 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-xkfg8" podStartSLOduration=10.263382846 podStartE2EDuration="10.263382846s" podCreationTimestamp="2025-10-02 07:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:48.259988217 +0000 UTC m=+149.291934494" watchObservedRunningTime="2025-10-02 07:18:48.263382846 +0000 UTC m=+149.295329133" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.281604 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2mgbj" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.363181 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.365684 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.865662468 +0000 UTC m=+149.897608755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.467628 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.468123 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:48.968102045 +0000 UTC m=+150.000048332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.515299 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.570854 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.571281 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.071261225 +0000 UTC m=+150.103207512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.636950 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ww8hs"] Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.703705 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.704677 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.204658255 +0000 UTC m=+150.236604542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.804842 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.805390 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.305352615 +0000 UTC m=+150.337298902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.805506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.805831 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.30581339 +0000 UTC m=+150.337759677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.906949 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:48 crc kubenswrapper[4960]: E1002 07:18:48.907381 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.407360978 +0000 UTC m=+150.439307265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.963585 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:48 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:48 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:48 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:48 crc kubenswrapper[4960]: I1002 07:18:48.963652 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.008553 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.008926 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.508912826 +0000 UTC m=+150.540859113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.055262 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvrdl"] Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.069465 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xwqkr"] Oct 02 07:18:49 crc kubenswrapper[4960]: W1002 07:18:49.079351 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b8d7ad4_3ab1_48af_a3f7_7c6783dd6ab2.slice/crio-2a641abc58f285753ff56cc54f758e6f22b252c28845f2ba1022a198ca0f7d15 WatchSource:0}: Error finding container 2a641abc58f285753ff56cc54f758e6f22b252c28845f2ba1022a198ca0f7d15: Status 404 returned error can't find the container with id 2a641abc58f285753ff56cc54f758e6f22b252c28845f2ba1022a198ca0f7d15 Oct 02 07:18:49 crc kubenswrapper[4960]: W1002 07:18:49.082909 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-67c97e48783f5486f622bfad856b3fb43d4ccaf41377a90725dae072ccfb2a04 WatchSource:0}: Error finding container 67c97e48783f5486f622bfad856b3fb43d4ccaf41377a90725dae072ccfb2a04: Status 404 returned error can't find the container with id 67c97e48783f5486f622bfad856b3fb43d4ccaf41377a90725dae072ccfb2a04 Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.088178 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqsbz"] Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.109547 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.109734 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.609685139 +0000 UTC m=+150.641631426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.109868 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.110221 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.610204957 +0000 UTC m=+150.642151244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: W1002 07:18:49.135998 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8438dbe_9de4_42cc_ae67_fc5368e55f58.slice/crio-94935e8aca5ea8ccc2061f4fcc89281f0020d0553805e76004bbcaf8f1799d25 WatchSource:0}: Error finding container 94935e8aca5ea8ccc2061f4fcc89281f0020d0553805e76004bbcaf8f1799d25: Status 404 returned error can't find the container with id 94935e8aca5ea8ccc2061f4fcc89281f0020d0553805e76004bbcaf8f1799d25 Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.149877 4960 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.210789 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.212474 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.712445947 +0000 UTC m=+150.744392234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.247968 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" event={"ID":"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327","Type":"ContainerStarted","Data":"01efd7360734862699afb1bc7120232d83d1e104e214a921345cebf97f33fae4"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.249278 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqsbz" event={"ID":"f8438dbe-9de4-42cc-ae67-fc5368e55f58","Type":"ContainerStarted","Data":"94935e8aca5ea8ccc2061f4fcc89281f0020d0553805e76004bbcaf8f1799d25"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.253567 4960 generic.go:334] "Generic (PLEG): container finished" podID="c816bb24-42ce-4155-906b-3817beeb1693" containerID="501ae48c041baa070ae17c0c11601d96a00ad5d9da513a3230c18a184758f091" exitCode=0 Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.253923 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ww8hs" event={"ID":"c816bb24-42ce-4155-906b-3817beeb1693","Type":"ContainerDied","Data":"501ae48c041baa070ae17c0c11601d96a00ad5d9da513a3230c18a184758f091"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.254011 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ww8hs" event={"ID":"c816bb24-42ce-4155-906b-3817beeb1693","Type":"ContainerStarted","Data":"196a70de0a5711c07e2881ad83077dcbe78bb92998f29b9818ed7229ddd89273"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.256749 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.263112 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwqkr" event={"ID":"92ef421d-9909-46b7-a040-4fecbc24ec8e","Type":"ContainerStarted","Data":"ef2c95875e716130e7c8bc5ee84c0f6686841f391846864535b53b8013cc70a4"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.270187 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6c0bd5bed398641489853073b5b93a5521604cc5c294287044f911ff3d9d8961"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.271136 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvrdl" event={"ID":"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2","Type":"ContainerStarted","Data":"2a641abc58f285753ff56cc54f758e6f22b252c28845f2ba1022a198ca0f7d15"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.272279 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f7751b62f23ed51b5ee0b02c02a3e6e91e4997cf2983e0b7f2892d96bc9b780c"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.276522 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"67c97e48783f5486f622bfad856b3fb43d4ccaf41377a90725dae072ccfb2a04"} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.314779 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.315563 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.815535505 +0000 UTC m=+150.847481792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.416550 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.416733 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.916697321 +0000 UTC m=+150.948643598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.417345 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.419684 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:18:49.919663317 +0000 UTC m=+150.951609604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8md6w" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.481258 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mm9j8"] Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.482266 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.484909 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.494029 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mm9j8"] Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.518787 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.519140 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:50.019119748 +0000 UTC m=+151.051066025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.549613 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.575170 4960 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T07:18:49.149933773Z","Handler":null,"Name":""} Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.578226 4960 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.578255 4960 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.620143 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b66wf\" (UniqueName: \"kubernetes.io/projected/7344a904-1c17-4585-88f4-36fa8449036a-kube-api-access-b66wf\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.620212 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.620433 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-utilities\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.620619 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-catalog-content\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.624020 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.624189 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.662158 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8md6w\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.722426 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.722505 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97a0f863-045c-4251-865d-327d1c1397e5-secret-volume\") pod \"97a0f863-045c-4251-865d-327d1c1397e5\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.722575 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djbnt\" (UniqueName: \"kubernetes.io/projected/97a0f863-045c-4251-865d-327d1c1397e5-kube-api-access-djbnt\") pod \"97a0f863-045c-4251-865d-327d1c1397e5\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.722666 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97a0f863-045c-4251-865d-327d1c1397e5-config-volume\") pod \"97a0f863-045c-4251-865d-327d1c1397e5\" (UID: \"97a0f863-045c-4251-865d-327d1c1397e5\") " Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.723005 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b66wf\" (UniqueName: \"kubernetes.io/projected/7344a904-1c17-4585-88f4-36fa8449036a-kube-api-access-b66wf\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.723089 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-utilities\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.723126 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-catalog-content\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.723541 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-utilities\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.723643 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-catalog-content\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.723676 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97a0f863-045c-4251-865d-327d1c1397e5-config-volume" (OuterVolumeSpecName: "config-volume") pod "97a0f863-045c-4251-865d-327d1c1397e5" (UID: "97a0f863-045c-4251-865d-327d1c1397e5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.726997 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.729816 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a0f863-045c-4251-865d-327d1c1397e5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "97a0f863-045c-4251-865d-327d1c1397e5" (UID: "97a0f863-045c-4251-865d-327d1c1397e5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.730528 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a0f863-045c-4251-865d-327d1c1397e5-kube-api-access-djbnt" (OuterVolumeSpecName: "kube-api-access-djbnt") pod "97a0f863-045c-4251-865d-327d1c1397e5" (UID: "97a0f863-045c-4251-865d-327d1c1397e5"). InnerVolumeSpecName "kube-api-access-djbnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.737755 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.745882 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b66wf\" (UniqueName: \"kubernetes.io/projected/7344a904-1c17-4585-88f4-36fa8449036a-kube-api-access-b66wf\") pod \"redhat-marketplace-mm9j8\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.828791 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97a0f863-045c-4251-865d-327d1c1397e5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.828834 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97a0f863-045c-4251-865d-327d1c1397e5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.828851 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djbnt\" (UniqueName: \"kubernetes.io/projected/97a0f863-045c-4251-865d-327d1c1397e5-kube-api-access-djbnt\") on node \"crc\" DevicePath \"\"" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.847592 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.883284 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s2h2d"] Oct 02 07:18:49 crc kubenswrapper[4960]: E1002 07:18:49.883583 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a0f863-045c-4251-865d-327d1c1397e5" containerName="collect-profiles" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.883600 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a0f863-045c-4251-865d-327d1c1397e5" containerName="collect-profiles" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.883753 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a0f863-045c-4251-865d-327d1c1397e5" containerName="collect-profiles" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.884605 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.910042 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2h2d"] Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.930255 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-catalog-content\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.930302 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrntp\" (UniqueName: \"kubernetes.io/projected/980cf347-0e45-456a-881e-3dbecb292223-kube-api-access-wrntp\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.930369 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-utilities\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.963628 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:49 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:49 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:49 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:49 crc kubenswrapper[4960]: I1002 07:18:49.963680 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.015235 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8md6w"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.031149 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-catalog-content\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.031200 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrntp\" (UniqueName: \"kubernetes.io/projected/980cf347-0e45-456a-881e-3dbecb292223-kube-api-access-wrntp\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.031270 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-utilities\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.031955 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-utilities\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.032537 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-catalog-content\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.060158 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrntp\" (UniqueName: \"kubernetes.io/projected/980cf347-0e45-456a-881e-3dbecb292223-kube-api-access-wrntp\") pod \"redhat-marketplace-s2h2d\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.129599 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mm9j8"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.214446 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.285713 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerID="0eec996b32119e354fc71e562e33b068625d7c090ab048578ef2287f7b087991" exitCode=0 Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.285826 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvrdl" event={"ID":"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2","Type":"ContainerDied","Data":"0eec996b32119e354fc71e562e33b068625d7c090ab048578ef2287f7b087991"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.291324 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6a6fa4c196fb8da4b4a034ba3182d5cfa5d21b5f87e5af3f48056f335bea2927"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.297012 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ff36e5fae2b89498a908dbac6bd9bc30a2968f0f934d99899ed7ce482bd75ce8"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.299632 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" event={"ID":"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e","Type":"ContainerStarted","Data":"b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.299682 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" event={"ID":"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e","Type":"ContainerStarted","Data":"ba1f051df8b4c74494d985b69e4f5da01a2d89f98e38a60a7b582ae878534ecb"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.300851 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.303205 4960 generic.go:334] "Generic (PLEG): container finished" podID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerID="8165153168f67aa9b2da0ae89ab609a2c0c059fd82cf657c760600c684b79d27" exitCode=0 Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.303501 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqsbz" event={"ID":"f8438dbe-9de4-42cc-ae67-fc5368e55f58","Type":"ContainerDied","Data":"8165153168f67aa9b2da0ae89ab609a2c0c059fd82cf657c760600c684b79d27"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.312813 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"393e3887f4835906fcd682aefa905b9aa8eb14e29d8a4b877778814aa937a09a"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.314282 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.342549 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.352763 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.354800 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn" event={"ID":"97a0f863-045c-4251-865d-327d1c1397e5","Type":"ContainerDied","Data":"16a8a614ddb1ed5dda660d698c8ec0b96b993c0ea3d7734839f4130a4dae7a45"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.354831 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16a8a614ddb1ed5dda660d698c8ec0b96b993c0ea3d7734839f4130a4dae7a45" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.357533 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" event={"ID":"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327","Type":"ContainerStarted","Data":"4ccfbb939a8b35f9bdb925b54e3e550429b530ac1661642e0ea71e40d7c7250f"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.357589 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" event={"ID":"d2ef18b3-68b3-4bc8-b8e0-2ff1ee1c1327","Type":"ContainerStarted","Data":"dd0aa4429391479315a5b499d2c7be5c2fc79c0f37e9a7ca688991a4735dc014"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.378148 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" podStartSLOduration=129.378122292 podStartE2EDuration="2m9.378122292s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:50.376959644 +0000 UTC m=+151.408905941" watchObservedRunningTime="2025-10-02 07:18:50.378122292 +0000 UTC m=+151.410068579" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.385246 4960 generic.go:334] "Generic (PLEG): container finished" podID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerID="2ade46fdcab2f06179e4da05134220550008ad221a723678ca743d166540f3a3" exitCode=0 Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.385417 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwqkr" event={"ID":"92ef421d-9909-46b7-a040-4fecbc24ec8e","Type":"ContainerDied","Data":"2ade46fdcab2f06179e4da05134220550008ad221a723678ca743d166540f3a3"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.409376 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mm9j8" event={"ID":"7344a904-1c17-4585-88f4-36fa8449036a","Type":"ContainerStarted","Data":"ed8a963dd18015e45281a8ff5d548adf784dc75e0405a0dcdce8c1f15d49c4b0"} Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.457667 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-nnbxf" podStartSLOduration=12.457640207 podStartE2EDuration="12.457640207s" podCreationTimestamp="2025-10-02 07:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:50.449206094 +0000 UTC m=+151.481152401" watchObservedRunningTime="2025-10-02 07:18:50.457640207 +0000 UTC m=+151.489586494" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.492678 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5gtzl"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.493646 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.498480 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.523317 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gtzl"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.560612 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.560661 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.594284 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-rc25f container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.594340 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rc25f" podUID="7e2be296-6def-442a-8771-f9333e891d14" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.594708 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-rc25f container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.594727 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rc25f" podUID="7e2be296-6def-442a-8771-f9333e891d14" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.605608 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.641455 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkb8h\" (UniqueName: \"kubernetes.io/projected/45228e94-0235-4947-8251-29b4ec09e509-kube-api-access-hkb8h\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.641584 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-utilities\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.641607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-catalog-content\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.670165 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.670926 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.673115 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.673480 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.698112 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2h2d"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.699014 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 07:18:50 crc kubenswrapper[4960]: W1002 07:18:50.712497 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod980cf347_0e45_456a_881e_3dbecb292223.slice/crio-bfaf84101356a3bc7ddc1084b3fcc83d13c828d4be225aa08fa5a530a5e9809b WatchSource:0}: Error finding container bfaf84101356a3bc7ddc1084b3fcc83d13c828d4be225aa08fa5a530a5e9809b: Status 404 returned error can't find the container with id bfaf84101356a3bc7ddc1084b3fcc83d13c828d4be225aa08fa5a530a5e9809b Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.743035 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-utilities\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.743097 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-catalog-content\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.743158 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkb8h\" (UniqueName: \"kubernetes.io/projected/45228e94-0235-4947-8251-29b4ec09e509-kube-api-access-hkb8h\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.745195 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-utilities\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.745446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-catalog-content\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.770803 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkb8h\" (UniqueName: \"kubernetes.io/projected/45228e94-0235-4947-8251-29b4ec09e509-kube-api-access-hkb8h\") pod \"redhat-operators-5gtzl\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.828947 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.844638 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.844707 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.882502 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rfmhs"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.887121 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.902083 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rfmhs"] Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.945601 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.945716 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.945799 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.963246 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:50 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:50 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:50 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.963312 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:50 crc kubenswrapper[4960]: I1002 07:18:50.973699 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.031545 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.051506 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-utilities\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.051582 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-catalog-content\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.051613 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72lqk\" (UniqueName: \"kubernetes.io/projected/4b58ae14-7933-418c-92e4-1eadb12dbde6-kube-api-access-72lqk\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.153330 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-utilities\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.153403 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-catalog-content\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.153435 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72lqk\" (UniqueName: \"kubernetes.io/projected/4b58ae14-7933-418c-92e4-1eadb12dbde6-kube-api-access-72lqk\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.154950 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-catalog-content\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.154961 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-utilities\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.177080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72lqk\" (UniqueName: \"kubernetes.io/projected/4b58ae14-7933-418c-92e4-1eadb12dbde6-kube-api-access-72lqk\") pod \"redhat-operators-rfmhs\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.213966 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gtzl"] Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.218017 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.296474 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.296900 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.298074 4960 patch_prober.go:28] interesting pod/console-f9d7485db-vz5gr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.298140 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vz5gr" podUID="a5911cf0-70ba-44a3-8c49-4cd66cca73b7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.429530 4960 generic.go:334] "Generic (PLEG): container finished" podID="980cf347-0e45-456a-881e-3dbecb292223" containerID="05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7" exitCode=0 Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.429612 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2h2d" event={"ID":"980cf347-0e45-456a-881e-3dbecb292223","Type":"ContainerDied","Data":"05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7"} Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.429648 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2h2d" event={"ID":"980cf347-0e45-456a-881e-3dbecb292223","Type":"ContainerStarted","Data":"bfaf84101356a3bc7ddc1084b3fcc83d13c828d4be225aa08fa5a530a5e9809b"} Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.432929 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gtzl" event={"ID":"45228e94-0235-4947-8251-29b4ec09e509","Type":"ContainerStarted","Data":"0cc06f44484d85211c73167e2d12b90cbc80aa84324097b6d94b825cf64b7ba2"} Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.435322 4960 generic.go:334] "Generic (PLEG): container finished" podID="7344a904-1c17-4585-88f4-36fa8449036a" containerID="133e8e154e0d9bd34c2680df5e7f90108494a871d7ba673f5ac6f59baeeed8ef" exitCode=0 Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.435864 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mm9j8" event={"ID":"7344a904-1c17-4585-88f4-36fa8449036a","Type":"ContainerDied","Data":"133e8e154e0d9bd34c2680df5e7f90108494a871d7ba673f5ac6f59baeeed8ef"} Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.443395 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-h8cqc" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.686252 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.777959 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.916553 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rfmhs"] Oct 02 07:18:51 crc kubenswrapper[4960]: W1002 07:18:51.937223 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b58ae14_7933_418c_92e4_1eadb12dbde6.slice/crio-73df7cb9721e3b6a3706936750c53f43c38f61f570d2a251c9e37b23ce17f9a5 WatchSource:0}: Error finding container 73df7cb9721e3b6a3706936750c53f43c38f61f570d2a251c9e37b23ce17f9a5: Status 404 returned error can't find the container with id 73df7cb9721e3b6a3706936750c53f43c38f61f570d2a251c9e37b23ce17f9a5 Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.959071 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.965598 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:51 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:51 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:51 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:51 crc kubenswrapper[4960]: I1002 07:18:51.965662 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.516601 4960 generic.go:334] "Generic (PLEG): container finished" podID="45228e94-0235-4947-8251-29b4ec09e509" containerID="51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08" exitCode=0 Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.517057 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gtzl" event={"ID":"45228e94-0235-4947-8251-29b4ec09e509","Type":"ContainerDied","Data":"51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08"} Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.520811 4960 generic.go:334] "Generic (PLEG): container finished" podID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerID="22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba" exitCode=0 Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.520876 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfmhs" event={"ID":"4b58ae14-7933-418c-92e4-1eadb12dbde6","Type":"ContainerDied","Data":"22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba"} Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.520912 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfmhs" event={"ID":"4b58ae14-7933-418c-92e4-1eadb12dbde6","Type":"ContainerStarted","Data":"73df7cb9721e3b6a3706936750c53f43c38f61f570d2a251c9e37b23ce17f9a5"} Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.523491 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e48aea6b-734b-4cbb-8568-3d716d8f0dc1","Type":"ContainerStarted","Data":"2445918570b97b6221bdd3b371dd38e3a20a9161133ceadb2be33a87981f71bd"} Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.963119 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:52 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:52 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:52 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:52 crc kubenswrapper[4960]: I1002 07:18:52.963177 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:53 crc kubenswrapper[4960]: I1002 07:18:53.539251 4960 generic.go:334] "Generic (PLEG): container finished" podID="e48aea6b-734b-4cbb-8568-3d716d8f0dc1" containerID="76b6d1a6369b52d1132e2fd4431d098d3b41c9453607e8a7cdff0bb3f3d3db51" exitCode=0 Oct 02 07:18:53 crc kubenswrapper[4960]: I1002 07:18:53.539304 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e48aea6b-734b-4cbb-8568-3d716d8f0dc1","Type":"ContainerDied","Data":"76b6d1a6369b52d1132e2fd4431d098d3b41c9453607e8a7cdff0bb3f3d3db51"} Oct 02 07:18:53 crc kubenswrapper[4960]: I1002 07:18:53.965612 4960 patch_prober.go:28] interesting pod/router-default-5444994796-9lf9c container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:18:53 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Oct 02 07:18:53 crc kubenswrapper[4960]: [+]process-running ok Oct 02 07:18:53 crc kubenswrapper[4960]: healthz check failed Oct 02 07:18:53 crc kubenswrapper[4960]: I1002 07:18:53.966151 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9lf9c" podUID="f6b93888-4d26-4500-ae07-aefff1dc0a7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.855522 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.856389 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.868686 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.868923 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.872652 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.944685 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/583693bf-61df-4d2a-a000-b4bdf20fc24a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.944744 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/583693bf-61df-4d2a-a000-b4bdf20fc24a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.964001 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.972118 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9lf9c" Oct 02 07:18:54 crc kubenswrapper[4960]: I1002 07:18:54.992553 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.052917 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/583693bf-61df-4d2a-a000-b4bdf20fc24a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.053112 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/583693bf-61df-4d2a-a000-b4bdf20fc24a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.053197 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/583693bf-61df-4d2a-a000-b4bdf20fc24a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.079738 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/583693bf-61df-4d2a-a000-b4bdf20fc24a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.156304 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kube-api-access\") pod \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.156794 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kubelet-dir\") pod \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\" (UID: \"e48aea6b-734b-4cbb-8568-3d716d8f0dc1\") " Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.156920 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e48aea6b-734b-4cbb-8568-3d716d8f0dc1" (UID: "e48aea6b-734b-4cbb-8568-3d716d8f0dc1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.157276 4960 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.163108 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e48aea6b-734b-4cbb-8568-3d716d8f0dc1" (UID: "e48aea6b-734b-4cbb-8568-3d716d8f0dc1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.182403 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.258614 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48aea6b-734b-4cbb-8568-3d716d8f0dc1-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.569913 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 07:18:55 crc kubenswrapper[4960]: W1002 07:18:55.590959 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod583693bf_61df_4d2a_a000_b4bdf20fc24a.slice/crio-ff3d34116ef95b0f64118db089b7533a0e2d7aac5d6b2141182ec1baff66bfb7 WatchSource:0}: Error finding container ff3d34116ef95b0f64118db089b7533a0e2d7aac5d6b2141182ec1baff66bfb7: Status 404 returned error can't find the container with id ff3d34116ef95b0f64118db089b7533a0e2d7aac5d6b2141182ec1baff66bfb7 Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.631682 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.631697 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e48aea6b-734b-4cbb-8568-3d716d8f0dc1","Type":"ContainerDied","Data":"2445918570b97b6221bdd3b371dd38e3a20a9161133ceadb2be33a87981f71bd"} Oct 02 07:18:55 crc kubenswrapper[4960]: I1002 07:18:55.631752 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2445918570b97b6221bdd3b371dd38e3a20a9161133ceadb2be33a87981f71bd" Oct 02 07:18:56 crc kubenswrapper[4960]: I1002 07:18:56.666255 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"583693bf-61df-4d2a-a000-b4bdf20fc24a","Type":"ContainerStarted","Data":"ff3d34116ef95b0f64118db089b7533a0e2d7aac5d6b2141182ec1baff66bfb7"} Oct 02 07:18:57 crc kubenswrapper[4960]: I1002 07:18:57.157988 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-xkfg8" Oct 02 07:18:57 crc kubenswrapper[4960]: I1002 07:18:57.682665 4960 generic.go:334] "Generic (PLEG): container finished" podID="583693bf-61df-4d2a-a000-b4bdf20fc24a" containerID="d8f2f4a8eb62ce1e59797a3af0c98307880e96ea223a81bde10ad32479f478c9" exitCode=0 Oct 02 07:18:57 crc kubenswrapper[4960]: I1002 07:18:57.682733 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"583693bf-61df-4d2a-a000-b4bdf20fc24a","Type":"ContainerDied","Data":"d8f2f4a8eb62ce1e59797a3af0c98307880e96ea223a81bde10ad32479f478c9"} Oct 02 07:18:59 crc kubenswrapper[4960]: I1002 07:18:59.150262 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:18:59 crc kubenswrapper[4960]: I1002 07:18:59.150852 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:19:00 crc kubenswrapper[4960]: I1002 07:19:00.608137 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-rc25f" Oct 02 07:19:01 crc kubenswrapper[4960]: I1002 07:19:01.307675 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:19:01 crc kubenswrapper[4960]: I1002 07:19:01.312751 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:19:04 crc kubenswrapper[4960]: I1002 07:19:04.269658 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:19:04 crc kubenswrapper[4960]: I1002 07:19:04.285774 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09886fa3-04f6-4e09-acc1-b983cbdfd6ba-metrics-certs\") pod \"network-metrics-daemon-hhvkr\" (UID: \"09886fa3-04f6-4e09-acc1-b983cbdfd6ba\") " pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:19:04 crc kubenswrapper[4960]: I1002 07:19:04.363832 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hhvkr" Oct 02 07:19:09 crc kubenswrapper[4960]: I1002 07:19:09.749466 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:19:22 crc kubenswrapper[4960]: I1002 07:19:22.325241 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hsbf5" Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.160487 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.218458 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/583693bf-61df-4d2a-a000-b4bdf20fc24a-kubelet-dir\") pod \"583693bf-61df-4d2a-a000-b4bdf20fc24a\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.218553 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/583693bf-61df-4d2a-a000-b4bdf20fc24a-kube-api-access\") pod \"583693bf-61df-4d2a-a000-b4bdf20fc24a\" (UID: \"583693bf-61df-4d2a-a000-b4bdf20fc24a\") " Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.220118 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/583693bf-61df-4d2a-a000-b4bdf20fc24a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "583693bf-61df-4d2a-a000-b4bdf20fc24a" (UID: "583693bf-61df-4d2a-a000-b4bdf20fc24a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.234668 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583693bf-61df-4d2a-a000-b4bdf20fc24a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "583693bf-61df-4d2a-a000-b4bdf20fc24a" (UID: "583693bf-61df-4d2a-a000-b4bdf20fc24a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.321168 4960 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/583693bf-61df-4d2a-a000-b4bdf20fc24a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.321213 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/583693bf-61df-4d2a-a000-b4bdf20fc24a-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.891755 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"583693bf-61df-4d2a-a000-b4bdf20fc24a","Type":"ContainerDied","Data":"ff3d34116ef95b0f64118db089b7533a0e2d7aac5d6b2141182ec1baff66bfb7"} Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.892228 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff3d34116ef95b0f64118db089b7533a0e2d7aac5d6b2141182ec1baff66bfb7" Oct 02 07:19:24 crc kubenswrapper[4960]: I1002 07:19:24.891824 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:25 crc kubenswrapper[4960]: E1002 07:19:25.740140 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 07:19:25 crc kubenswrapper[4960]: E1002 07:19:25.740437 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-97sc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xvrdl_openshift-marketplace(1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:25 crc kubenswrapper[4960]: E1002 07:19:25.741912 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xvrdl" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" Oct 02 07:19:25 crc kubenswrapper[4960]: E1002 07:19:25.745664 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 07:19:25 crc kubenswrapper[4960]: E1002 07:19:25.745896 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-58pvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xwqkr_openshift-marketplace(92ef421d-9909-46b7-a040-4fecbc24ec8e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:25 crc kubenswrapper[4960]: E1002 07:19:25.747063 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xwqkr" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" Oct 02 07:19:26 crc kubenswrapper[4960]: E1002 07:19:26.104547 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xwqkr" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" Oct 02 07:19:26 crc kubenswrapper[4960]: E1002 07:19:26.104701 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xvrdl" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" Oct 02 07:19:27 crc kubenswrapper[4960]: E1002 07:19:27.536558 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 07:19:27 crc kubenswrapper[4960]: E1002 07:19:27.537373 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d2gzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fqsbz_openshift-marketplace(f8438dbe-9de4-42cc-ae67-fc5368e55f58): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:27 crc kubenswrapper[4960]: E1002 07:19:27.538665 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fqsbz" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" Oct 02 07:19:27 crc kubenswrapper[4960]: E1002 07:19:27.547077 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 07:19:27 crc kubenswrapper[4960]: E1002 07:19:27.547333 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-prb5t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ww8hs_openshift-marketplace(c816bb24-42ce-4155-906b-3817beeb1693): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:27 crc kubenswrapper[4960]: E1002 07:19:27.548579 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-ww8hs" podUID="c816bb24-42ce-4155-906b-3817beeb1693" Oct 02 07:19:27 crc kubenswrapper[4960]: I1002 07:19:27.762549 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:29 crc kubenswrapper[4960]: I1002 07:19:29.150776 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:19:29 crc kubenswrapper[4960]: I1002 07:19:29.151440 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.672889 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fqsbz" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.673279 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ww8hs" podUID="c816bb24-42ce-4155-906b-3817beeb1693" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.759357 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.759635 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-72lqk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rfmhs_openshift-marketplace(4b58ae14-7933-418c-92e4-1eadb12dbde6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.760916 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rfmhs" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.773080 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.773281 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hkb8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5gtzl_openshift-marketplace(45228e94-0235-4947-8251-29b4ec09e509): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:30 crc kubenswrapper[4960]: E1002 07:19:30.774422 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5gtzl" podUID="45228e94-0235-4947-8251-29b4ec09e509" Oct 02 07:19:33 crc kubenswrapper[4960]: E1002 07:19:33.688503 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rfmhs" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" Oct 02 07:19:33 crc kubenswrapper[4960]: E1002 07:19:33.689110 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5gtzl" podUID="45228e94-0235-4947-8251-29b4ec09e509" Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.313809 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.314471 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrntp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-s2h2d_openshift-marketplace(980cf347-0e45-456a-881e-3dbecb292223): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.316170 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-s2h2d" podUID="980cf347-0e45-456a-881e-3dbecb292223" Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.316243 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.316505 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b66wf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mm9j8_openshift-marketplace(7344a904-1c17-4585-88f4-36fa8449036a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.318594 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mm9j8" podUID="7344a904-1c17-4585-88f4-36fa8449036a" Oct 02 07:19:34 crc kubenswrapper[4960]: I1002 07:19:34.651376 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hhvkr"] Oct 02 07:19:34 crc kubenswrapper[4960]: I1002 07:19:34.950348 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" event={"ID":"09886fa3-04f6-4e09-acc1-b983cbdfd6ba","Type":"ContainerStarted","Data":"4304328abb4fdbded31310e051dccfa335f0a2f792fe107d05a8a8001997bb4d"} Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.954277 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-s2h2d" podUID="980cf347-0e45-456a-881e-3dbecb292223" Oct 02 07:19:34 crc kubenswrapper[4960]: E1002 07:19:34.954372 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mm9j8" podUID="7344a904-1c17-4585-88f4-36fa8449036a" Oct 02 07:19:35 crc kubenswrapper[4960]: I1002 07:19:35.961010 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" event={"ID":"09886fa3-04f6-4e09-acc1-b983cbdfd6ba","Type":"ContainerStarted","Data":"4d3855041c3fd4e6b0bf849542eedbb25d1cf0fd113d374d42020e039922ef62"} Oct 02 07:19:35 crc kubenswrapper[4960]: I1002 07:19:35.961088 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hhvkr" event={"ID":"09886fa3-04f6-4e09-acc1-b983cbdfd6ba","Type":"ContainerStarted","Data":"9a79bdd2992d725693d3b81ba72c1be4eb6b2a2efe1b85c27da20ba2ecbf942c"} Oct 02 07:19:35 crc kubenswrapper[4960]: I1002 07:19:35.996508 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hhvkr" podStartSLOduration=174.996472937 podStartE2EDuration="2m54.996472937s" podCreationTimestamp="2025-10-02 07:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:35.991222556 +0000 UTC m=+197.023168883" watchObservedRunningTime="2025-10-02 07:19:35.996472937 +0000 UTC m=+197.028419254" Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.053472 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerID="b60be8d3c6a980b1aac8ef4d9f63d412cdeedb3fd3f7eec9d0bb8adc26fab743" exitCode=0 Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.054090 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvrdl" event={"ID":"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2","Type":"ContainerDied","Data":"b60be8d3c6a980b1aac8ef4d9f63d412cdeedb3fd3f7eec9d0bb8adc26fab743"} Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.064673 4960 generic.go:334] "Generic (PLEG): container finished" podID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerID="a8c98672490cad99ad1b12e2582972ff68629711c4e9def0fc0ad9ef38bda8a9" exitCode=0 Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.064735 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqsbz" event={"ID":"f8438dbe-9de4-42cc-ae67-fc5368e55f58","Type":"ContainerDied","Data":"a8c98672490cad99ad1b12e2582972ff68629711c4e9def0fc0ad9ef38bda8a9"} Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.067551 4960 generic.go:334] "Generic (PLEG): container finished" podID="c816bb24-42ce-4155-906b-3817beeb1693" containerID="8647960ef3f8f72c32dfb0b867c4f5ace5861d93f69701bc8b83c3f9e456a5c9" exitCode=0 Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.067599 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ww8hs" event={"ID":"c816bb24-42ce-4155-906b-3817beeb1693","Type":"ContainerDied","Data":"8647960ef3f8f72c32dfb0b867c4f5ace5861d93f69701bc8b83c3f9e456a5c9"} Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.071126 4960 generic.go:334] "Generic (PLEG): container finished" podID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerID="50ead06bcb416b6b0987618becc2eb756751761fdbe1d93850d6a94d1990bb47" exitCode=0 Oct 02 07:19:48 crc kubenswrapper[4960]: I1002 07:19:48.071182 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwqkr" event={"ID":"92ef421d-9909-46b7-a040-4fecbc24ec8e","Type":"ContainerDied","Data":"50ead06bcb416b6b0987618becc2eb756751761fdbe1d93850d6a94d1990bb47"} Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.083888 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvrdl" event={"ID":"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2","Type":"ContainerStarted","Data":"54c528ac52b07f759d3b7a1a0ffd46698d354ca864255d57fafda2e09d2781f9"} Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.086122 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqsbz" event={"ID":"f8438dbe-9de4-42cc-ae67-fc5368e55f58","Type":"ContainerStarted","Data":"5f26c98c28633747bf15b97ae9cb008edacb6e1a577e585bdfceb87ec0723c5a"} Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.088045 4960 generic.go:334] "Generic (PLEG): container finished" podID="980cf347-0e45-456a-881e-3dbecb292223" containerID="a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6" exitCode=0 Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.088110 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2h2d" event={"ID":"980cf347-0e45-456a-881e-3dbecb292223","Type":"ContainerDied","Data":"a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6"} Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.091310 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ww8hs" event={"ID":"c816bb24-42ce-4155-906b-3817beeb1693","Type":"ContainerStarted","Data":"3dec038ee78ccbacd56c3b30976c3f28e8ec36a69bde9c54a19579598154d992"} Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.096454 4960 generic.go:334] "Generic (PLEG): container finished" podID="45228e94-0235-4947-8251-29b4ec09e509" containerID="bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502" exitCode=0 Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.096575 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gtzl" event={"ID":"45228e94-0235-4947-8251-29b4ec09e509","Type":"ContainerDied","Data":"bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502"} Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.102217 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwqkr" event={"ID":"92ef421d-9909-46b7-a040-4fecbc24ec8e","Type":"ContainerStarted","Data":"4e4622677e977b3561c568c97392b2acef6f2178fc337584b0c951dcac0bae8f"} Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.127127 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xvrdl" podStartSLOduration=3.842499459 podStartE2EDuration="1m2.127099474s" podCreationTimestamp="2025-10-02 07:18:47 +0000 UTC" firstStartedPulling="2025-10-02 07:18:50.287597331 +0000 UTC m=+151.319543618" lastFinishedPulling="2025-10-02 07:19:48.572197336 +0000 UTC m=+209.604143633" observedRunningTime="2025-10-02 07:19:49.125946024 +0000 UTC m=+210.157892311" watchObservedRunningTime="2025-10-02 07:19:49.127099474 +0000 UTC m=+210.159045761" Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.150143 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xwqkr" podStartSLOduration=4.0622143059999996 podStartE2EDuration="1m2.150123481s" podCreationTimestamp="2025-10-02 07:18:47 +0000 UTC" firstStartedPulling="2025-10-02 07:18:50.388738636 +0000 UTC m=+151.420684913" lastFinishedPulling="2025-10-02 07:19:48.476647801 +0000 UTC m=+209.508594088" observedRunningTime="2025-10-02 07:19:49.144831478 +0000 UTC m=+210.176777765" watchObservedRunningTime="2025-10-02 07:19:49.150123481 +0000 UTC m=+210.182069768" Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.170357 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fqsbz" podStartSLOduration=3.971257976 podStartE2EDuration="1m2.17033533s" podCreationTimestamp="2025-10-02 07:18:47 +0000 UTC" firstStartedPulling="2025-10-02 07:18:50.309142859 +0000 UTC m=+151.341089146" lastFinishedPulling="2025-10-02 07:19:48.508220213 +0000 UTC m=+209.540166500" observedRunningTime="2025-10-02 07:19:49.166524548 +0000 UTC m=+210.198470835" watchObservedRunningTime="2025-10-02 07:19:49.17033533 +0000 UTC m=+210.202281627" Oct 02 07:19:49 crc kubenswrapper[4960]: I1002 07:19:49.216916 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ww8hs" podStartSLOduration=2.81274716 podStartE2EDuration="1m2.21689319s" podCreationTimestamp="2025-10-02 07:18:47 +0000 UTC" firstStartedPulling="2025-10-02 07:18:49.255600195 +0000 UTC m=+150.287546482" lastFinishedPulling="2025-10-02 07:19:48.659746215 +0000 UTC m=+209.691692512" observedRunningTime="2025-10-02 07:19:49.19577841 +0000 UTC m=+210.227724697" watchObservedRunningTime="2025-10-02 07:19:49.21689319 +0000 UTC m=+210.248839477" Oct 02 07:19:50 crc kubenswrapper[4960]: I1002 07:19:50.110957 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gtzl" event={"ID":"45228e94-0235-4947-8251-29b4ec09e509","Type":"ContainerStarted","Data":"570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1"} Oct 02 07:19:50 crc kubenswrapper[4960]: I1002 07:19:50.144263 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5gtzl" podStartSLOduration=3.162628317 podStartE2EDuration="1m0.144237805s" podCreationTimestamp="2025-10-02 07:18:50 +0000 UTC" firstStartedPulling="2025-10-02 07:18:52.521727533 +0000 UTC m=+153.553673820" lastFinishedPulling="2025-10-02 07:19:49.503337011 +0000 UTC m=+210.535283308" observedRunningTime="2025-10-02 07:19:50.141910514 +0000 UTC m=+211.173856791" watchObservedRunningTime="2025-10-02 07:19:50.144237805 +0000 UTC m=+211.176184092" Oct 02 07:19:50 crc kubenswrapper[4960]: I1002 07:19:50.829684 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:19:50 crc kubenswrapper[4960]: I1002 07:19:50.830289 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:19:51 crc kubenswrapper[4960]: I1002 07:19:51.122834 4960 generic.go:334] "Generic (PLEG): container finished" podID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerID="7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218" exitCode=0 Oct 02 07:19:51 crc kubenswrapper[4960]: I1002 07:19:51.122908 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfmhs" event={"ID":"4b58ae14-7933-418c-92e4-1eadb12dbde6","Type":"ContainerDied","Data":"7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218"} Oct 02 07:19:51 crc kubenswrapper[4960]: I1002 07:19:51.126776 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2h2d" event={"ID":"980cf347-0e45-456a-881e-3dbecb292223","Type":"ContainerStarted","Data":"69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101"} Oct 02 07:19:51 crc kubenswrapper[4960]: I1002 07:19:51.179380 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s2h2d" podStartSLOduration=3.111250838 podStartE2EDuration="1m2.179356497s" podCreationTimestamp="2025-10-02 07:18:49 +0000 UTC" firstStartedPulling="2025-10-02 07:18:51.431495681 +0000 UTC m=+152.463441968" lastFinishedPulling="2025-10-02 07:19:50.49960134 +0000 UTC m=+211.531547627" observedRunningTime="2025-10-02 07:19:51.175822874 +0000 UTC m=+212.207769161" watchObservedRunningTime="2025-10-02 07:19:51.179356497 +0000 UTC m=+212.211302784" Oct 02 07:19:51 crc kubenswrapper[4960]: I1002 07:19:51.982085 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5gtzl" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="registry-server" probeResult="failure" output=< Oct 02 07:19:51 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 07:19:51 crc kubenswrapper[4960]: > Oct 02 07:19:52 crc kubenswrapper[4960]: I1002 07:19:52.154839 4960 generic.go:334] "Generic (PLEG): container finished" podID="7344a904-1c17-4585-88f4-36fa8449036a" containerID="0849300e54a20513ec33027f0cc5eeb50b964598584dd344f9bbe6565b5df51b" exitCode=0 Oct 02 07:19:52 crc kubenswrapper[4960]: I1002 07:19:52.154931 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mm9j8" event={"ID":"7344a904-1c17-4585-88f4-36fa8449036a","Type":"ContainerDied","Data":"0849300e54a20513ec33027f0cc5eeb50b964598584dd344f9bbe6565b5df51b"} Oct 02 07:19:52 crc kubenswrapper[4960]: I1002 07:19:52.158141 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfmhs" event={"ID":"4b58ae14-7933-418c-92e4-1eadb12dbde6","Type":"ContainerStarted","Data":"965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733"} Oct 02 07:19:52 crc kubenswrapper[4960]: I1002 07:19:52.201282 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rfmhs" podStartSLOduration=3.118335485 podStartE2EDuration="1m2.201261022s" podCreationTimestamp="2025-10-02 07:18:50 +0000 UTC" firstStartedPulling="2025-10-02 07:18:52.526015072 +0000 UTC m=+153.557961359" lastFinishedPulling="2025-10-02 07:19:51.608940609 +0000 UTC m=+212.640886896" observedRunningTime="2025-10-02 07:19:52.196911302 +0000 UTC m=+213.228857589" watchObservedRunningTime="2025-10-02 07:19:52.201261022 +0000 UTC m=+213.233207309" Oct 02 07:19:53 crc kubenswrapper[4960]: I1002 07:19:53.165322 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mm9j8" event={"ID":"7344a904-1c17-4585-88f4-36fa8449036a","Type":"ContainerStarted","Data":"9558a2541694eff2a0e4d689a303d5e4dee813de3907660ad6ef0bee3f2219c8"} Oct 02 07:19:53 crc kubenswrapper[4960]: I1002 07:19:53.186032 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mm9j8" podStartSLOduration=2.060453329 podStartE2EDuration="1m4.186015132s" podCreationTimestamp="2025-10-02 07:18:49 +0000 UTC" firstStartedPulling="2025-10-02 07:18:50.414269393 +0000 UTC m=+151.446215680" lastFinishedPulling="2025-10-02 07:19:52.539831186 +0000 UTC m=+213.571777483" observedRunningTime="2025-10-02 07:19:53.183548656 +0000 UTC m=+214.215494943" watchObservedRunningTime="2025-10-02 07:19:53.186015132 +0000 UTC m=+214.217961419" Oct 02 07:19:57 crc kubenswrapper[4960]: I1002 07:19:57.643100 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:19:57 crc kubenswrapper[4960]: I1002 07:19:57.643561 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:19:57 crc kubenswrapper[4960]: I1002 07:19:57.695341 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:19:57 crc kubenswrapper[4960]: I1002 07:19:57.805584 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:19:57 crc kubenswrapper[4960]: I1002 07:19:57.805638 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:19:57 crc kubenswrapper[4960]: I1002 07:19:57.859918 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.064836 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.064906 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.104289 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.232755 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.244967 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.255727 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.515718 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.515812 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:19:58 crc kubenswrapper[4960]: I1002 07:19:58.554669 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.149635 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.149710 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.149765 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.150434 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.150584 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25" gracePeriod=600 Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.239163 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.848356 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.848442 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:19:59 crc kubenswrapper[4960]: I1002 07:19:59.905727 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.215677 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.215783 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.251139 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.275841 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.390384 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvrdl"] Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.390709 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xvrdl" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="registry-server" containerID="cri-o://54c528ac52b07f759d3b7a1a0ffd46698d354ca864255d57fafda2e09d2781f9" gracePeriod=2 Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.602468 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fr4q7"] Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.887664 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:20:00 crc kubenswrapper[4960]: I1002 07:20:00.952032 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.011655 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqsbz"] Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.210237 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25" exitCode=0 Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.210696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25"} Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.213249 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerID="54c528ac52b07f759d3b7a1a0ffd46698d354ca864255d57fafda2e09d2781f9" exitCode=0 Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.213451 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvrdl" event={"ID":"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2","Type":"ContainerDied","Data":"54c528ac52b07f759d3b7a1a0ffd46698d354ca864255d57fafda2e09d2781f9"} Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.213774 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fqsbz" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="registry-server" containerID="cri-o://5f26c98c28633747bf15b97ae9cb008edacb6e1a577e585bdfceb87ec0723c5a" gracePeriod=2 Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.219613 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.220786 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.254055 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.267618 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.614665 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.646611 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-utilities\") pod \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.646700 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97sc7\" (UniqueName: \"kubernetes.io/projected/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-kube-api-access-97sc7\") pod \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.646759 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-catalog-content\") pod \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\" (UID: \"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2\") " Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.647668 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-utilities" (OuterVolumeSpecName: "utilities") pod "1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" (UID: "1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.653504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-kube-api-access-97sc7" (OuterVolumeSpecName: "kube-api-access-97sc7") pod "1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" (UID: "1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2"). InnerVolumeSpecName "kube-api-access-97sc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.710827 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" (UID: "1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.748965 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.749037 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97sc7\" (UniqueName: \"kubernetes.io/projected/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-kube-api-access-97sc7\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:01 crc kubenswrapper[4960]: I1002 07:20:01.749052 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.227548 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvrdl" event={"ID":"1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2","Type":"ContainerDied","Data":"2a641abc58f285753ff56cc54f758e6f22b252c28845f2ba1022a198ca0f7d15"} Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.227617 4960 scope.go:117] "RemoveContainer" containerID="54c528ac52b07f759d3b7a1a0ffd46698d354ca864255d57fafda2e09d2781f9" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.227680 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvrdl" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.233259 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"e05a2b35123183dd07d94c3c5aa895e60ba41f5e31b710be48e94c02186f6ea1"} Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.236996 4960 generic.go:334] "Generic (PLEG): container finished" podID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerID="5f26c98c28633747bf15b97ae9cb008edacb6e1a577e585bdfceb87ec0723c5a" exitCode=0 Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.237056 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqsbz" event={"ID":"f8438dbe-9de4-42cc-ae67-fc5368e55f58","Type":"ContainerDied","Data":"5f26c98c28633747bf15b97ae9cb008edacb6e1a577e585bdfceb87ec0723c5a"} Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.251395 4960 scope.go:117] "RemoveContainer" containerID="b60be8d3c6a980b1aac8ef4d9f63d412cdeedb3fd3f7eec9d0bb8adc26fab743" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.281057 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvrdl"] Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.286390 4960 scope.go:117] "RemoveContainer" containerID="0eec996b32119e354fc71e562e33b068625d7c090ab048578ef2287f7b087991" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.286724 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xvrdl"] Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.305605 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.337722 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" path="/var/lib/kubelet/pods/1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2/volumes" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.503125 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.560629 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2gzl\" (UniqueName: \"kubernetes.io/projected/f8438dbe-9de4-42cc-ae67-fc5368e55f58-kube-api-access-d2gzl\") pod \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.560728 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-catalog-content\") pod \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.560756 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-utilities\") pod \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\" (UID: \"f8438dbe-9de4-42cc-ae67-fc5368e55f58\") " Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.562384 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-utilities" (OuterVolumeSpecName: "utilities") pod "f8438dbe-9de4-42cc-ae67-fc5368e55f58" (UID: "f8438dbe-9de4-42cc-ae67-fc5368e55f58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.571877 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8438dbe-9de4-42cc-ae67-fc5368e55f58-kube-api-access-d2gzl" (OuterVolumeSpecName: "kube-api-access-d2gzl") pod "f8438dbe-9de4-42cc-ae67-fc5368e55f58" (UID: "f8438dbe-9de4-42cc-ae67-fc5368e55f58"). InnerVolumeSpecName "kube-api-access-d2gzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.610876 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8438dbe-9de4-42cc-ae67-fc5368e55f58" (UID: "f8438dbe-9de4-42cc-ae67-fc5368e55f58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.661935 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2gzl\" (UniqueName: \"kubernetes.io/projected/f8438dbe-9de4-42cc-ae67-fc5368e55f58-kube-api-access-d2gzl\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.662041 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.662057 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8438dbe-9de4-42cc-ae67-fc5368e55f58-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:02 crc kubenswrapper[4960]: I1002 07:20:02.789606 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2h2d"] Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.247749 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqsbz" event={"ID":"f8438dbe-9de4-42cc-ae67-fc5368e55f58","Type":"ContainerDied","Data":"94935e8aca5ea8ccc2061f4fcc89281f0020d0553805e76004bbcaf8f1799d25"} Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.248101 4960 scope.go:117] "RemoveContainer" containerID="5f26c98c28633747bf15b97ae9cb008edacb6e1a577e585bdfceb87ec0723c5a" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.248002 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqsbz" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.249235 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s2h2d" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="registry-server" containerID="cri-o://69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101" gracePeriod=2 Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.268924 4960 scope.go:117] "RemoveContainer" containerID="a8c98672490cad99ad1b12e2582972ff68629711c4e9def0fc0ad9ef38bda8a9" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.282655 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqsbz"] Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.285910 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fqsbz"] Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.310080 4960 scope.go:117] "RemoveContainer" containerID="8165153168f67aa9b2da0ae89ab609a2c0c059fd82cf657c760600c684b79d27" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.678260 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.778659 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrntp\" (UniqueName: \"kubernetes.io/projected/980cf347-0e45-456a-881e-3dbecb292223-kube-api-access-wrntp\") pod \"980cf347-0e45-456a-881e-3dbecb292223\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.778847 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-catalog-content\") pod \"980cf347-0e45-456a-881e-3dbecb292223\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.778927 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-utilities\") pod \"980cf347-0e45-456a-881e-3dbecb292223\" (UID: \"980cf347-0e45-456a-881e-3dbecb292223\") " Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.786532 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/980cf347-0e45-456a-881e-3dbecb292223-kube-api-access-wrntp" (OuterVolumeSpecName: "kube-api-access-wrntp") pod "980cf347-0e45-456a-881e-3dbecb292223" (UID: "980cf347-0e45-456a-881e-3dbecb292223"). InnerVolumeSpecName "kube-api-access-wrntp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.787071 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-utilities" (OuterVolumeSpecName: "utilities") pod "980cf347-0e45-456a-881e-3dbecb292223" (UID: "980cf347-0e45-456a-881e-3dbecb292223"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.801959 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "980cf347-0e45-456a-881e-3dbecb292223" (UID: "980cf347-0e45-456a-881e-3dbecb292223"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.880327 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrntp\" (UniqueName: \"kubernetes.io/projected/980cf347-0e45-456a-881e-3dbecb292223-kube-api-access-wrntp\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.880378 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:03 crc kubenswrapper[4960]: I1002 07:20:03.880388 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980cf347-0e45-456a-881e-3dbecb292223-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.257882 4960 generic.go:334] "Generic (PLEG): container finished" podID="980cf347-0e45-456a-881e-3dbecb292223" containerID="69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101" exitCode=0 Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.257947 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2h2d" event={"ID":"980cf347-0e45-456a-881e-3dbecb292223","Type":"ContainerDied","Data":"69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101"} Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.258711 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2h2d" event={"ID":"980cf347-0e45-456a-881e-3dbecb292223","Type":"ContainerDied","Data":"bfaf84101356a3bc7ddc1084b3fcc83d13c828d4be225aa08fa5a530a5e9809b"} Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.258020 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2h2d" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.258835 4960 scope.go:117] "RemoveContainer" containerID="69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.276866 4960 scope.go:117] "RemoveContainer" containerID="a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.308336 4960 scope.go:117] "RemoveContainer" containerID="05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.310242 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2h2d"] Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.313955 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2h2d"] Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.334004 4960 scope.go:117] "RemoveContainer" containerID="69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101" Oct 02 07:20:04 crc kubenswrapper[4960]: E1002 07:20:04.335719 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101\": container with ID starting with 69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101 not found: ID does not exist" containerID="69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.335754 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101"} err="failed to get container status \"69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101\": rpc error: code = NotFound desc = could not find container \"69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101\": container with ID starting with 69adabdc41545212b2ffb4ed5f37926cdacfd17e0f61249d54fdaebdb7916101 not found: ID does not exist" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.335783 4960 scope.go:117] "RemoveContainer" containerID="a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6" Oct 02 07:20:04 crc kubenswrapper[4960]: E1002 07:20:04.336290 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6\": container with ID starting with a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6 not found: ID does not exist" containerID="a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.336351 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6"} err="failed to get container status \"a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6\": rpc error: code = NotFound desc = could not find container \"a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6\": container with ID starting with a4367f3453bc28b9ae95fadf7454a2f1f4d171372b5c76bc30606038eaab95c6 not found: ID does not exist" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.336387 4960 scope.go:117] "RemoveContainer" containerID="05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7" Oct 02 07:20:04 crc kubenswrapper[4960]: E1002 07:20:04.336962 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7\": container with ID starting with 05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7 not found: ID does not exist" containerID="05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.337003 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7"} err="failed to get container status \"05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7\": rpc error: code = NotFound desc = could not find container \"05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7\": container with ID starting with 05594a1965c79165c9b59a5039f52d88268357c9ad8b28639909a61ad87f08d7 not found: ID does not exist" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.337212 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="980cf347-0e45-456a-881e-3dbecb292223" path="/var/lib/kubelet/pods/980cf347-0e45-456a-881e-3dbecb292223/volumes" Oct 02 07:20:04 crc kubenswrapper[4960]: I1002 07:20:04.337843 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" path="/var/lib/kubelet/pods/f8438dbe-9de4-42cc-ae67-fc5368e55f58/volumes" Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.186659 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rfmhs"] Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.186925 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rfmhs" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="registry-server" containerID="cri-o://965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733" gracePeriod=2 Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.561167 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.707902 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-utilities\") pod \"4b58ae14-7933-418c-92e4-1eadb12dbde6\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.708131 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72lqk\" (UniqueName: \"kubernetes.io/projected/4b58ae14-7933-418c-92e4-1eadb12dbde6-kube-api-access-72lqk\") pod \"4b58ae14-7933-418c-92e4-1eadb12dbde6\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.708186 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-catalog-content\") pod \"4b58ae14-7933-418c-92e4-1eadb12dbde6\" (UID: \"4b58ae14-7933-418c-92e4-1eadb12dbde6\") " Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.711700 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-utilities" (OuterVolumeSpecName: "utilities") pod "4b58ae14-7933-418c-92e4-1eadb12dbde6" (UID: "4b58ae14-7933-418c-92e4-1eadb12dbde6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.718138 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b58ae14-7933-418c-92e4-1eadb12dbde6-kube-api-access-72lqk" (OuterVolumeSpecName: "kube-api-access-72lqk") pod "4b58ae14-7933-418c-92e4-1eadb12dbde6" (UID: "4b58ae14-7933-418c-92e4-1eadb12dbde6"). InnerVolumeSpecName "kube-api-access-72lqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.809739 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72lqk\" (UniqueName: \"kubernetes.io/projected/4b58ae14-7933-418c-92e4-1eadb12dbde6-kube-api-access-72lqk\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.809782 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.823997 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b58ae14-7933-418c-92e4-1eadb12dbde6" (UID: "4b58ae14-7933-418c-92e4-1eadb12dbde6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:05 crc kubenswrapper[4960]: I1002 07:20:05.911593 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b58ae14-7933-418c-92e4-1eadb12dbde6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.278943 4960 generic.go:334] "Generic (PLEG): container finished" podID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerID="965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733" exitCode=0 Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.279252 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfmhs" event={"ID":"4b58ae14-7933-418c-92e4-1eadb12dbde6","Type":"ContainerDied","Data":"965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733"} Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.279283 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfmhs" event={"ID":"4b58ae14-7933-418c-92e4-1eadb12dbde6","Type":"ContainerDied","Data":"73df7cb9721e3b6a3706936750c53f43c38f61f570d2a251c9e37b23ce17f9a5"} Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.279304 4960 scope.go:117] "RemoveContainer" containerID="965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.279405 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rfmhs" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.299548 4960 scope.go:117] "RemoveContainer" containerID="7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.316068 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rfmhs"] Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.322713 4960 scope.go:117] "RemoveContainer" containerID="22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.324483 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rfmhs"] Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.339781 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" path="/var/lib/kubelet/pods/4b58ae14-7933-418c-92e4-1eadb12dbde6/volumes" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.344037 4960 scope.go:117] "RemoveContainer" containerID="965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733" Oct 02 07:20:06 crc kubenswrapper[4960]: E1002 07:20:06.344794 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733\": container with ID starting with 965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733 not found: ID does not exist" containerID="965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.344846 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733"} err="failed to get container status \"965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733\": rpc error: code = NotFound desc = could not find container \"965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733\": container with ID starting with 965bf4d0a4beca5fcff27ac78bb97c390fa6cb0394711b701ed859faf07b4733 not found: ID does not exist" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.344880 4960 scope.go:117] "RemoveContainer" containerID="7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218" Oct 02 07:20:06 crc kubenswrapper[4960]: E1002 07:20:06.345405 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218\": container with ID starting with 7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218 not found: ID does not exist" containerID="7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.345437 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218"} err="failed to get container status \"7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218\": rpc error: code = NotFound desc = could not find container \"7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218\": container with ID starting with 7ba06ec45581cbe0ee165d7dfad3948d7c544ddf89453f04b0e3fec2962b4218 not found: ID does not exist" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.345469 4960 scope.go:117] "RemoveContainer" containerID="22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba" Oct 02 07:20:06 crc kubenswrapper[4960]: E1002 07:20:06.345959 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba\": container with ID starting with 22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba not found: ID does not exist" containerID="22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba" Oct 02 07:20:06 crc kubenswrapper[4960]: I1002 07:20:06.346115 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba"} err="failed to get container status \"22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba\": rpc error: code = NotFound desc = could not find container \"22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba\": container with ID starting with 22d08e6d7fd7d80b399f24834d73d8720827b862aff938a7e3376ac2de13c0ba not found: ID does not exist" Oct 02 07:20:25 crc kubenswrapper[4960]: I1002 07:20:25.641650 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" podUID="be880348-4fad-4b89-8e00-6dd336636f76" containerName="oauth-openshift" containerID="cri-o://3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0" gracePeriod=15 Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.118598 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.165610 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps"] Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.165906 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.165925 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.165936 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.165945 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.165961 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.165971 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166002 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166011 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166066 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be880348-4fad-4b89-8e00-6dd336636f76" containerName="oauth-openshift" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166200 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="be880348-4fad-4b89-8e00-6dd336636f76" containerName="oauth-openshift" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166214 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166222 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166231 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583693bf-61df-4d2a-a000-b4bdf20fc24a" containerName="pruner" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166242 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="583693bf-61df-4d2a-a000-b4bdf20fc24a" containerName="pruner" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166254 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166262 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166275 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48aea6b-734b-4cbb-8568-3d716d8f0dc1" containerName="pruner" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166283 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48aea6b-734b-4cbb-8568-3d716d8f0dc1" containerName="pruner" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166293 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166301 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166313 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166322 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166333 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166341 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166352 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166360 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166375 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166384 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="extract-content" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.166393 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166401 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="extract-utilities" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166519 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="583693bf-61df-4d2a-a000-b4bdf20fc24a" containerName="pruner" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166538 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8438dbe-9de4-42cc-ae67-fc5368e55f58" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166550 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="be880348-4fad-4b89-8e00-6dd336636f76" containerName="oauth-openshift" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166562 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48aea6b-734b-4cbb-8568-3d716d8f0dc1" containerName="pruner" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166573 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="980cf347-0e45-456a-881e-3dbecb292223" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166583 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b8d7ad4-3ab1-48af-a3f7-7c6783dd6ab2" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.166594 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b58ae14-7933-418c-92e4-1eadb12dbde6" containerName="registry-server" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.167086 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.177819 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps"] Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207377 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-provider-selection\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207443 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-login\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207469 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-serving-cert\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207491 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-service-ca\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207518 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-router-certs\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207543 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be880348-4fad-4b89-8e00-6dd336636f76-audit-dir\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207567 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-cliconfig\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207585 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-trusted-ca-bundle\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207603 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-idp-0-file-data\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207637 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-session\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207656 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-ocp-branding-template\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207685 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-error\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207726 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-audit-policies\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207768 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5dqd\" (UniqueName: \"kubernetes.io/projected/be880348-4fad-4b89-8e00-6dd336636f76-kube-api-access-l5dqd\") pod \"be880348-4fad-4b89-8e00-6dd336636f76\" (UID: \"be880348-4fad-4b89-8e00-6dd336636f76\") " Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207912 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/828eda15-afd4-495a-af15-cd7baf5f349e-audit-dir\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207957 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdnl4\" (UniqueName: \"kubernetes.io/projected/828eda15-afd4-495a-af15-cd7baf5f349e-kube-api-access-vdnl4\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.207998 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208024 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208041 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208061 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208081 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208107 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-audit-policies\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208126 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208148 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208167 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208192 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.208219 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.211158 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be880348-4fad-4b89-8e00-6dd336636f76-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.212242 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.212350 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.213230 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.214296 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.217239 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.218486 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.222400 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.222992 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be880348-4fad-4b89-8e00-6dd336636f76-kube-api-access-l5dqd" (OuterVolumeSpecName: "kube-api-access-l5dqd") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "kube-api-access-l5dqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.223322 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.226575 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.227659 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.227845 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.228238 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "be880348-4fad-4b89-8e00-6dd336636f76" (UID: "be880348-4fad-4b89-8e00-6dd336636f76"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308694 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308748 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308775 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308804 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308836 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/828eda15-afd4-495a-af15-cd7baf5f349e-audit-dir\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308857 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308876 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdnl4\" (UniqueName: \"kubernetes.io/projected/828eda15-afd4-495a-af15-cd7baf5f349e-kube-api-access-vdnl4\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308913 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308930 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308950 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.308969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309013 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-audit-policies\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309034 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309076 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309089 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309109 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309410 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309430 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309445 4960 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be880348-4fad-4b89-8e00-6dd336636f76-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309464 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309482 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309494 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309506 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309522 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309537 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/be880348-4fad-4b89-8e00-6dd336636f76-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309549 4960 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be880348-4fad-4b89-8e00-6dd336636f76-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.309567 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5dqd\" (UniqueName: \"kubernetes.io/projected/be880348-4fad-4b89-8e00-6dd336636f76-kube-api-access-l5dqd\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.310013 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/828eda15-afd4-495a-af15-cd7baf5f349e-audit-dir\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.310066 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.310581 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.310736 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.311232 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/828eda15-afd4-495a-af15-cd7baf5f349e-audit-policies\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.314045 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.314515 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.314600 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.314801 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.315110 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.315227 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.315709 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.315957 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/828eda15-afd4-495a-af15-cd7baf5f349e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.331025 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdnl4\" (UniqueName: \"kubernetes.io/projected/828eda15-afd4-495a-af15-cd7baf5f349e-kube-api-access-vdnl4\") pod \"oauth-openshift-57bcd9fbb-8b7ps\" (UID: \"828eda15-afd4-495a-af15-cd7baf5f349e\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.396118 4960 generic.go:334] "Generic (PLEG): container finished" podID="be880348-4fad-4b89-8e00-6dd336636f76" containerID="3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0" exitCode=0 Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.396173 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" event={"ID":"be880348-4fad-4b89-8e00-6dd336636f76","Type":"ContainerDied","Data":"3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0"} Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.396202 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" event={"ID":"be880348-4fad-4b89-8e00-6dd336636f76","Type":"ContainerDied","Data":"38ad047e4cfcd97f201f66cbc3e153a827a013cad4e773e2c2044893f3cb7274"} Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.396223 4960 scope.go:117] "RemoveContainer" containerID="3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.396341 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fr4q7" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.418539 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fr4q7"] Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.421126 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fr4q7"] Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.430566 4960 scope.go:117] "RemoveContainer" containerID="3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0" Oct 02 07:20:26 crc kubenswrapper[4960]: E1002 07:20:26.430962 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0\": container with ID starting with 3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0 not found: ID does not exist" containerID="3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.431020 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0"} err="failed to get container status \"3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0\": rpc error: code = NotFound desc = could not find container \"3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0\": container with ID starting with 3ddc9e49b5c9542c7143a6a141c6aca6baa21ea94862d8f2fc14c015fbf750e0 not found: ID does not exist" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.492827 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:26 crc kubenswrapper[4960]: I1002 07:20:26.920689 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps"] Oct 02 07:20:27 crc kubenswrapper[4960]: I1002 07:20:27.403226 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" event={"ID":"828eda15-afd4-495a-af15-cd7baf5f349e","Type":"ContainerStarted","Data":"b51f0e9eeff8b35d992a379cab0a641d849cfcd66c4e54f4f21e91f9f89fd73b"} Oct 02 07:20:27 crc kubenswrapper[4960]: I1002 07:20:27.403598 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" event={"ID":"828eda15-afd4-495a-af15-cd7baf5f349e","Type":"ContainerStarted","Data":"82e0b2df15b09b93b856dc6cf1af2130f2967d707506efa38ce2cc8d68f81c96"} Oct 02 07:20:27 crc kubenswrapper[4960]: I1002 07:20:27.403615 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:27 crc kubenswrapper[4960]: I1002 07:20:27.430311 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" podStartSLOduration=27.430287725 podStartE2EDuration="27.430287725s" podCreationTimestamp="2025-10-02 07:20:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:20:27.424725856 +0000 UTC m=+248.456672143" watchObservedRunningTime="2025-10-02 07:20:27.430287725 +0000 UTC m=+248.462234012" Oct 02 07:20:27 crc kubenswrapper[4960]: I1002 07:20:27.846916 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57bcd9fbb-8b7ps" Oct 02 07:20:28 crc kubenswrapper[4960]: I1002 07:20:28.337636 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be880348-4fad-4b89-8e00-6dd336636f76" path="/var/lib/kubelet/pods/be880348-4fad-4b89-8e00-6dd336636f76/volumes" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.128998 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ww8hs"] Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.130021 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ww8hs" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="registry-server" containerID="cri-o://3dec038ee78ccbacd56c3b30976c3f28e8ec36a69bde9c54a19579598154d992" gracePeriod=30 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.141852 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xwqkr"] Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.142315 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xwqkr" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="registry-server" containerID="cri-o://4e4622677e977b3561c568c97392b2acef6f2178fc337584b0c951dcac0bae8f" gracePeriod=30 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.156410 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-96r9w"] Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.156659 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" podUID="0122bd5f-ff81-4468-aac9-32f277f12e68" containerName="marketplace-operator" containerID="cri-o://59a97a7f46fb0f0e705e6caee2ef204e07004914c23cc4722493d121b465c9e9" gracePeriod=30 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.166289 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mm9j8"] Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.166625 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mm9j8" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="registry-server" containerID="cri-o://9558a2541694eff2a0e4d689a303d5e4dee813de3907660ad6ef0bee3f2219c8" gracePeriod=30 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.178520 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gtzl"] Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.179566 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5gtzl" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="registry-server" containerID="cri-o://570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1" gracePeriod=30 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.215928 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbbzq"] Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.217308 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.222461 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbbzq"] Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.262633 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82lm7\" (UniqueName: \"kubernetes.io/projected/8a54100f-6654-4909-b7cc-ce0b74ded3f5-kube-api-access-82lm7\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.263115 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a54100f-6654-4909-b7cc-ce0b74ded3f5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.263165 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8a54100f-6654-4909-b7cc-ce0b74ded3f5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.370519 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8a54100f-6654-4909-b7cc-ce0b74ded3f5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.370724 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82lm7\" (UniqueName: \"kubernetes.io/projected/8a54100f-6654-4909-b7cc-ce0b74ded3f5-kube-api-access-82lm7\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.370764 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a54100f-6654-4909-b7cc-ce0b74ded3f5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.374729 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a54100f-6654-4909-b7cc-ce0b74ded3f5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.378827 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8a54100f-6654-4909-b7cc-ce0b74ded3f5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.405942 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82lm7\" (UniqueName: \"kubernetes.io/projected/8a54100f-6654-4909-b7cc-ce0b74ded3f5-kube-api-access-82lm7\") pod \"marketplace-operator-79b997595-tbbzq\" (UID: \"8a54100f-6654-4909-b7cc-ce0b74ded3f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.504779 4960 generic.go:334] "Generic (PLEG): container finished" podID="7344a904-1c17-4585-88f4-36fa8449036a" containerID="9558a2541694eff2a0e4d689a303d5e4dee813de3907660ad6ef0bee3f2219c8" exitCode=0 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.504835 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mm9j8" event={"ID":"7344a904-1c17-4585-88f4-36fa8449036a","Type":"ContainerDied","Data":"9558a2541694eff2a0e4d689a303d5e4dee813de3907660ad6ef0bee3f2219c8"} Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.506929 4960 generic.go:334] "Generic (PLEG): container finished" podID="0122bd5f-ff81-4468-aac9-32f277f12e68" containerID="59a97a7f46fb0f0e705e6caee2ef204e07004914c23cc4722493d121b465c9e9" exitCode=0 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.506987 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" event={"ID":"0122bd5f-ff81-4468-aac9-32f277f12e68","Type":"ContainerDied","Data":"59a97a7f46fb0f0e705e6caee2ef204e07004914c23cc4722493d121b465c9e9"} Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.510234 4960 generic.go:334] "Generic (PLEG): container finished" podID="c816bb24-42ce-4155-906b-3817beeb1693" containerID="3dec038ee78ccbacd56c3b30976c3f28e8ec36a69bde9c54a19579598154d992" exitCode=0 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.510273 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ww8hs" event={"ID":"c816bb24-42ce-4155-906b-3817beeb1693","Type":"ContainerDied","Data":"3dec038ee78ccbacd56c3b30976c3f28e8ec36a69bde9c54a19579598154d992"} Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.516549 4960 generic.go:334] "Generic (PLEG): container finished" podID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerID="4e4622677e977b3561c568c97392b2acef6f2178fc337584b0c951dcac0bae8f" exitCode=0 Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.516623 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwqkr" event={"ID":"92ef421d-9909-46b7-a040-4fecbc24ec8e","Type":"ContainerDied","Data":"4e4622677e977b3561c568c97392b2acef6f2178fc337584b0c951dcac0bae8f"} Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.583532 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.598819 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.674313 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prb5t\" (UniqueName: \"kubernetes.io/projected/c816bb24-42ce-4155-906b-3817beeb1693-kube-api-access-prb5t\") pod \"c816bb24-42ce-4155-906b-3817beeb1693\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.674817 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-utilities\") pod \"c816bb24-42ce-4155-906b-3817beeb1693\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.674895 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-catalog-content\") pod \"c816bb24-42ce-4155-906b-3817beeb1693\" (UID: \"c816bb24-42ce-4155-906b-3817beeb1693\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.677895 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-utilities" (OuterVolumeSpecName: "utilities") pod "c816bb24-42ce-4155-906b-3817beeb1693" (UID: "c816bb24-42ce-4155-906b-3817beeb1693"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.683457 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c816bb24-42ce-4155-906b-3817beeb1693-kube-api-access-prb5t" (OuterVolumeSpecName: "kube-api-access-prb5t") pod "c816bb24-42ce-4155-906b-3817beeb1693" (UID: "c816bb24-42ce-4155-906b-3817beeb1693"). InnerVolumeSpecName "kube-api-access-prb5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.714420 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.714542 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.720279 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.732171 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c816bb24-42ce-4155-906b-3817beeb1693" (UID: "c816bb24-42ce-4155-906b-3817beeb1693"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.776737 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-operator-metrics\") pod \"0122bd5f-ff81-4468-aac9-32f277f12e68\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.776876 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-utilities\") pod \"92ef421d-9909-46b7-a040-4fecbc24ec8e\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.777668 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-utilities" (OuterVolumeSpecName: "utilities") pod "92ef421d-9909-46b7-a040-4fecbc24ec8e" (UID: "92ef421d-9909-46b7-a040-4fecbc24ec8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.778475 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b66wf\" (UniqueName: \"kubernetes.io/projected/7344a904-1c17-4585-88f4-36fa8449036a-kube-api-access-b66wf\") pod \"7344a904-1c17-4585-88f4-36fa8449036a\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.778507 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58pvm\" (UniqueName: \"kubernetes.io/projected/92ef421d-9909-46b7-a040-4fecbc24ec8e-kube-api-access-58pvm\") pod \"92ef421d-9909-46b7-a040-4fecbc24ec8e\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.778557 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-utilities\") pod \"7344a904-1c17-4585-88f4-36fa8449036a\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.778589 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-catalog-content\") pod \"7344a904-1c17-4585-88f4-36fa8449036a\" (UID: \"7344a904-1c17-4585-88f4-36fa8449036a\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.778632 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd9dz\" (UniqueName: \"kubernetes.io/projected/0122bd5f-ff81-4468-aac9-32f277f12e68-kube-api-access-cd9dz\") pod \"0122bd5f-ff81-4468-aac9-32f277f12e68\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.778658 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-catalog-content\") pod \"92ef421d-9909-46b7-a040-4fecbc24ec8e\" (UID: \"92ef421d-9909-46b7-a040-4fecbc24ec8e\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.778794 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-trusted-ca\") pod \"0122bd5f-ff81-4468-aac9-32f277f12e68\" (UID: \"0122bd5f-ff81-4468-aac9-32f277f12e68\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.779087 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.779101 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prb5t\" (UniqueName: \"kubernetes.io/projected/c816bb24-42ce-4155-906b-3817beeb1693-kube-api-access-prb5t\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.779114 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c816bb24-42ce-4155-906b-3817beeb1693-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.779124 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.779665 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-utilities" (OuterVolumeSpecName: "utilities") pod "7344a904-1c17-4585-88f4-36fa8449036a" (UID: "7344a904-1c17-4585-88f4-36fa8449036a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.780943 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0122bd5f-ff81-4468-aac9-32f277f12e68" (UID: "0122bd5f-ff81-4468-aac9-32f277f12e68"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.782295 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0122bd5f-ff81-4468-aac9-32f277f12e68" (UID: "0122bd5f-ff81-4468-aac9-32f277f12e68"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.784145 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7344a904-1c17-4585-88f4-36fa8449036a-kube-api-access-b66wf" (OuterVolumeSpecName: "kube-api-access-b66wf") pod "7344a904-1c17-4585-88f4-36fa8449036a" (UID: "7344a904-1c17-4585-88f4-36fa8449036a"). InnerVolumeSpecName "kube-api-access-b66wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.791550 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0122bd5f-ff81-4468-aac9-32f277f12e68-kube-api-access-cd9dz" (OuterVolumeSpecName: "kube-api-access-cd9dz") pod "0122bd5f-ff81-4468-aac9-32f277f12e68" (UID: "0122bd5f-ff81-4468-aac9-32f277f12e68"). InnerVolumeSpecName "kube-api-access-cd9dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.793964 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ef421d-9909-46b7-a040-4fecbc24ec8e-kube-api-access-58pvm" (OuterVolumeSpecName: "kube-api-access-58pvm") pod "92ef421d-9909-46b7-a040-4fecbc24ec8e" (UID: "92ef421d-9909-46b7-a040-4fecbc24ec8e"). InnerVolumeSpecName "kube-api-access-58pvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.817299 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7344a904-1c17-4585-88f4-36fa8449036a" (UID: "7344a904-1c17-4585-88f4-36fa8449036a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.851706 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92ef421d-9909-46b7-a040-4fecbc24ec8e" (UID: "92ef421d-9909-46b7-a040-4fecbc24ec8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.861138 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880277 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-utilities\") pod \"45228e94-0235-4947-8251-29b4ec09e509\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880350 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkb8h\" (UniqueName: \"kubernetes.io/projected/45228e94-0235-4947-8251-29b4ec09e509-kube-api-access-hkb8h\") pod \"45228e94-0235-4947-8251-29b4ec09e509\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880457 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-catalog-content\") pod \"45228e94-0235-4947-8251-29b4ec09e509\" (UID: \"45228e94-0235-4947-8251-29b4ec09e509\") " Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880746 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b66wf\" (UniqueName: \"kubernetes.io/projected/7344a904-1c17-4585-88f4-36fa8449036a-kube-api-access-b66wf\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880782 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58pvm\" (UniqueName: \"kubernetes.io/projected/92ef421d-9909-46b7-a040-4fecbc24ec8e-kube-api-access-58pvm\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880801 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880820 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7344a904-1c17-4585-88f4-36fa8449036a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880832 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd9dz\" (UniqueName: \"kubernetes.io/projected/0122bd5f-ff81-4468-aac9-32f277f12e68-kube-api-access-cd9dz\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880844 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ef421d-9909-46b7-a040-4fecbc24ec8e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880855 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.880867 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0122bd5f-ff81-4468-aac9-32f277f12e68-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.881591 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-utilities" (OuterVolumeSpecName: "utilities") pod "45228e94-0235-4947-8251-29b4ec09e509" (UID: "45228e94-0235-4947-8251-29b4ec09e509"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.891013 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45228e94-0235-4947-8251-29b4ec09e509-kube-api-access-hkb8h" (OuterVolumeSpecName: "kube-api-access-hkb8h") pod "45228e94-0235-4947-8251-29b4ec09e509" (UID: "45228e94-0235-4947-8251-29b4ec09e509"). InnerVolumeSpecName "kube-api-access-hkb8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.981793 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45228e94-0235-4947-8251-29b4ec09e509" (UID: "45228e94-0235-4947-8251-29b4ec09e509"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.982625 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.982648 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45228e94-0235-4947-8251-29b4ec09e509-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4960]: I1002 07:20:44.982660 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkb8h\" (UniqueName: \"kubernetes.io/projected/45228e94-0235-4947-8251-29b4ec09e509-kube-api-access-hkb8h\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.070557 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbbzq"] Oct 02 07:20:45 crc kubenswrapper[4960]: W1002 07:20:45.077933 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a54100f_6654_4909_b7cc_ce0b74ded3f5.slice/crio-154ff137550abb96b68600d4b55d033be42772315af74f8f7eed22f87ece2599 WatchSource:0}: Error finding container 154ff137550abb96b68600d4b55d033be42772315af74f8f7eed22f87ece2599: Status 404 returned error can't find the container with id 154ff137550abb96b68600d4b55d033be42772315af74f8f7eed22f87ece2599 Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.526352 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.526399 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-96r9w" event={"ID":"0122bd5f-ff81-4468-aac9-32f277f12e68","Type":"ContainerDied","Data":"a59dfca4fb7b9d21d63f71b2f420e14eeb8966a957cc7cae1e1b31bbac21feed"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.526496 4960 scope.go:117] "RemoveContainer" containerID="59a97a7f46fb0f0e705e6caee2ef204e07004914c23cc4722493d121b465c9e9" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.528604 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" event={"ID":"8a54100f-6654-4909-b7cc-ce0b74ded3f5","Type":"ContainerStarted","Data":"db74d0af65a9c8b16cbf976ac41e24f5fa8e3790517c525f5d09a0f053a88467"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.528670 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" event={"ID":"8a54100f-6654-4909-b7cc-ce0b74ded3f5","Type":"ContainerStarted","Data":"154ff137550abb96b68600d4b55d033be42772315af74f8f7eed22f87ece2599"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.528836 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.531767 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ww8hs" event={"ID":"c816bb24-42ce-4155-906b-3817beeb1693","Type":"ContainerDied","Data":"196a70de0a5711c07e2881ad83077dcbe78bb92998f29b9818ed7229ddd89273"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.531806 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ww8hs" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.534142 4960 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tbbzq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.534198 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" podUID="8a54100f-6654-4909-b7cc-ce0b74ded3f5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.546608 4960 generic.go:334] "Generic (PLEG): container finished" podID="45228e94-0235-4947-8251-29b4ec09e509" containerID="570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1" exitCode=0 Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.546810 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gtzl" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.546857 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gtzl" event={"ID":"45228e94-0235-4947-8251-29b4ec09e509","Type":"ContainerDied","Data":"570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.546964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gtzl" event={"ID":"45228e94-0235-4947-8251-29b4ec09e509","Type":"ContainerDied","Data":"0cc06f44484d85211c73167e2d12b90cbc80aa84324097b6d94b825cf64b7ba2"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.559117 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwqkr" event={"ID":"92ef421d-9909-46b7-a040-4fecbc24ec8e","Type":"ContainerDied","Data":"ef2c95875e716130e7c8bc5ee84c0f6686841f391846864535b53b8013cc70a4"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.559243 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwqkr" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.563047 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mm9j8" event={"ID":"7344a904-1c17-4585-88f4-36fa8449036a","Type":"ContainerDied","Data":"ed8a963dd18015e45281a8ff5d548adf784dc75e0405a0dcdce8c1f15d49c4b0"} Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.563151 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mm9j8" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.564841 4960 scope.go:117] "RemoveContainer" containerID="3dec038ee78ccbacd56c3b30976c3f28e8ec36a69bde9c54a19579598154d992" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.577792 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" podStartSLOduration=1.577715773 podStartE2EDuration="1.577715773s" podCreationTimestamp="2025-10-02 07:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:20:45.566280602 +0000 UTC m=+266.598226899" watchObservedRunningTime="2025-10-02 07:20:45.577715773 +0000 UTC m=+266.609662060" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.590358 4960 scope.go:117] "RemoveContainer" containerID="8647960ef3f8f72c32dfb0b867c4f5ace5861d93f69701bc8b83c3f9e456a5c9" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.608160 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ww8hs"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.612158 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ww8hs"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.615358 4960 scope.go:117] "RemoveContainer" containerID="501ae48c041baa070ae17c0c11601d96a00ad5d9da513a3230c18a184758f091" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.626650 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gtzl"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.633405 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5gtzl"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.645623 4960 scope.go:117] "RemoveContainer" containerID="570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.647069 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-96r9w"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.649412 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-96r9w"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.657525 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xwqkr"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.660485 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xwqkr"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.664449 4960 scope.go:117] "RemoveContainer" containerID="bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.669823 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mm9j8"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.681662 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mm9j8"] Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.683340 4960 scope.go:117] "RemoveContainer" containerID="51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.707109 4960 scope.go:117] "RemoveContainer" containerID="570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1" Oct 02 07:20:45 crc kubenswrapper[4960]: E1002 07:20:45.707935 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1\": container with ID starting with 570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1 not found: ID does not exist" containerID="570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.707968 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1"} err="failed to get container status \"570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1\": rpc error: code = NotFound desc = could not find container \"570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1\": container with ID starting with 570a32c9202845eb7018908e7edc78dd28950a24905ddec9a9a58748e579b0a1 not found: ID does not exist" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.708013 4960 scope.go:117] "RemoveContainer" containerID="bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502" Oct 02 07:20:45 crc kubenswrapper[4960]: E1002 07:20:45.708516 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502\": container with ID starting with bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502 not found: ID does not exist" containerID="bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.708639 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502"} err="failed to get container status \"bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502\": rpc error: code = NotFound desc = could not find container \"bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502\": container with ID starting with bbf446e4457f89f90171d5e13587ee92338c8e9a8ef2b69c729feb1589642502 not found: ID does not exist" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.708752 4960 scope.go:117] "RemoveContainer" containerID="51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08" Oct 02 07:20:45 crc kubenswrapper[4960]: E1002 07:20:45.709375 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08\": container with ID starting with 51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08 not found: ID does not exist" containerID="51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.709502 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08"} err="failed to get container status \"51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08\": rpc error: code = NotFound desc = could not find container \"51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08\": container with ID starting with 51dbde4a2cf0a9ff7454c1076bd109bcaf9f747caba9305083990719bc8e8a08 not found: ID does not exist" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.709595 4960 scope.go:117] "RemoveContainer" containerID="4e4622677e977b3561c568c97392b2acef6f2178fc337584b0c951dcac0bae8f" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.726302 4960 scope.go:117] "RemoveContainer" containerID="50ead06bcb416b6b0987618becc2eb756751761fdbe1d93850d6a94d1990bb47" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.750603 4960 scope.go:117] "RemoveContainer" containerID="2ade46fdcab2f06179e4da05134220550008ad221a723678ca743d166540f3a3" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.768652 4960 scope.go:117] "RemoveContainer" containerID="9558a2541694eff2a0e4d689a303d5e4dee813de3907660ad6ef0bee3f2219c8" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.786186 4960 scope.go:117] "RemoveContainer" containerID="0849300e54a20513ec33027f0cc5eeb50b964598584dd344f9bbe6565b5df51b" Oct 02 07:20:45 crc kubenswrapper[4960]: I1002 07:20:45.801231 4960 scope.go:117] "RemoveContainer" containerID="133e8e154e0d9bd34c2680df5e7f90108494a871d7ba673f5ac6f59baeeed8ef" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.340499 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0122bd5f-ff81-4468-aac9-32f277f12e68" path="/var/lib/kubelet/pods/0122bd5f-ff81-4468-aac9-32f277f12e68/volumes" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.345708 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45228e94-0235-4947-8251-29b4ec09e509" path="/var/lib/kubelet/pods/45228e94-0235-4947-8251-29b4ec09e509/volumes" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.347269 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7344a904-1c17-4585-88f4-36fa8449036a" path="/var/lib/kubelet/pods/7344a904-1c17-4585-88f4-36fa8449036a/volumes" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.353744 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" path="/var/lib/kubelet/pods/92ef421d-9909-46b7-a040-4fecbc24ec8e/volumes" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.355756 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c816bb24-42ce-4155-906b-3817beeb1693" path="/var/lib/kubelet/pods/c816bb24-42ce-4155-906b-3817beeb1693/volumes" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.357869 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rq7lt"] Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360467 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360515 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360554 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360574 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360596 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360613 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360641 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360678 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360696 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360716 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360738 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0122bd5f-ff81-4468-aac9-32f277f12e68" containerName="marketplace-operator" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360756 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0122bd5f-ff81-4468-aac9-32f277f12e68" containerName="marketplace-operator" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360784 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360801 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360818 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360831 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="extract-utilities" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360853 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360869 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360886 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360900 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360919 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360932 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.360947 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.360961 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: E1002 07:20:46.361012 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.361026 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="extract-content" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.361201 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ef421d-9909-46b7-a040-4fecbc24ec8e" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.361238 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c816bb24-42ce-4155-906b-3817beeb1693" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.361255 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0122bd5f-ff81-4468-aac9-32f277f12e68" containerName="marketplace-operator" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.361271 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="45228e94-0235-4947-8251-29b4ec09e509" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.361295 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7344a904-1c17-4585-88f4-36fa8449036a" containerName="registry-server" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.363661 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.366917 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.368961 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rq7lt"] Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.409289 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-catalog-content\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.409372 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-utilities\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.409410 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv76t\" (UniqueName: \"kubernetes.io/projected/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-kube-api-access-kv76t\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.511194 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-catalog-content\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.511257 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-utilities\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.511279 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv76t\" (UniqueName: \"kubernetes.io/projected/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-kube-api-access-kv76t\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.512044 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-utilities\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.512225 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-catalog-content\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.556169 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nw7n6"] Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.557208 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.560335 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.563236 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv76t\" (UniqueName: \"kubernetes.io/projected/ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7-kube-api-access-kv76t\") pod \"redhat-marketplace-rq7lt\" (UID: \"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7\") " pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.580058 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nw7n6"] Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.583745 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tbbzq" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.617679 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e601cc-5a5e-4630-9693-a7f6c42174d9-catalog-content\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.617737 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z649j\" (UniqueName: \"kubernetes.io/projected/95e601cc-5a5e-4630-9693-a7f6c42174d9-kube-api-access-z649j\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.618016 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e601cc-5a5e-4630-9693-a7f6c42174d9-utilities\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.684535 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.719725 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z649j\" (UniqueName: \"kubernetes.io/projected/95e601cc-5a5e-4630-9693-a7f6c42174d9-kube-api-access-z649j\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.719827 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e601cc-5a5e-4630-9693-a7f6c42174d9-utilities\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.719866 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e601cc-5a5e-4630-9693-a7f6c42174d9-catalog-content\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.720382 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e601cc-5a5e-4630-9693-a7f6c42174d9-catalog-content\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.721098 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e601cc-5a5e-4630-9693-a7f6c42174d9-utilities\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.737767 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z649j\" (UniqueName: \"kubernetes.io/projected/95e601cc-5a5e-4630-9693-a7f6c42174d9-kube-api-access-z649j\") pod \"redhat-operators-nw7n6\" (UID: \"95e601cc-5a5e-4630-9693-a7f6c42174d9\") " pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.893902 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rq7lt"] Oct 02 07:20:46 crc kubenswrapper[4960]: I1002 07:20:46.898393 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:46 crc kubenswrapper[4960]: W1002 07:20:46.903160 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae6a6e63_ef7b_4fdf_bf06_994e0651fdb7.slice/crio-a1aef7be8288c6a51ba22443c9ea6422dba7d2479088894b9b32a79e0570f2b7 WatchSource:0}: Error finding container a1aef7be8288c6a51ba22443c9ea6422dba7d2479088894b9b32a79e0570f2b7: Status 404 returned error can't find the container with id a1aef7be8288c6a51ba22443c9ea6422dba7d2479088894b9b32a79e0570f2b7 Oct 02 07:20:47 crc kubenswrapper[4960]: I1002 07:20:47.104403 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nw7n6"] Oct 02 07:20:47 crc kubenswrapper[4960]: W1002 07:20:47.170842 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95e601cc_5a5e_4630_9693_a7f6c42174d9.slice/crio-7b3eb753aef35a69c0dfc49640a8ee306dbac1256319ce3a38144823c0d745ef WatchSource:0}: Error finding container 7b3eb753aef35a69c0dfc49640a8ee306dbac1256319ce3a38144823c0d745ef: Status 404 returned error can't find the container with id 7b3eb753aef35a69c0dfc49640a8ee306dbac1256319ce3a38144823c0d745ef Oct 02 07:20:47 crc kubenswrapper[4960]: I1002 07:20:47.591131 4960 generic.go:334] "Generic (PLEG): container finished" podID="95e601cc-5a5e-4630-9693-a7f6c42174d9" containerID="8ed7b099b317e8f104c824495803cc9b89df405c3a26ab2281568a380e025c7b" exitCode=0 Oct 02 07:20:47 crc kubenswrapper[4960]: I1002 07:20:47.591230 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw7n6" event={"ID":"95e601cc-5a5e-4630-9693-a7f6c42174d9","Type":"ContainerDied","Data":"8ed7b099b317e8f104c824495803cc9b89df405c3a26ab2281568a380e025c7b"} Oct 02 07:20:47 crc kubenswrapper[4960]: I1002 07:20:47.592855 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw7n6" event={"ID":"95e601cc-5a5e-4630-9693-a7f6c42174d9","Type":"ContainerStarted","Data":"7b3eb753aef35a69c0dfc49640a8ee306dbac1256319ce3a38144823c0d745ef"} Oct 02 07:20:47 crc kubenswrapper[4960]: I1002 07:20:47.598657 4960 generic.go:334] "Generic (PLEG): container finished" podID="ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7" containerID="8a8604f5c063ead8cc23172a41862def9936f2494d1a6a7afc9ce89800b6a3d4" exitCode=0 Oct 02 07:20:47 crc kubenswrapper[4960]: I1002 07:20:47.598914 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq7lt" event={"ID":"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7","Type":"ContainerDied","Data":"8a8604f5c063ead8cc23172a41862def9936f2494d1a6a7afc9ce89800b6a3d4"} Oct 02 07:20:47 crc kubenswrapper[4960]: I1002 07:20:47.599032 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq7lt" event={"ID":"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7","Type":"ContainerStarted","Data":"a1aef7be8288c6a51ba22443c9ea6422dba7d2479088894b9b32a79e0570f2b7"} Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.606559 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw7n6" event={"ID":"95e601cc-5a5e-4630-9693-a7f6c42174d9","Type":"ContainerStarted","Data":"bafc79410f5dc41ae394b51e799c7dde479a61398919f26d7f69a4605f38b73b"} Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.609603 4960 generic.go:334] "Generic (PLEG): container finished" podID="ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7" containerID="2de211893f5be288b79b4a1fd765d0fa259d160fcb6f4cf42935dafe8605d5cd" exitCode=0 Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.609649 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq7lt" event={"ID":"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7","Type":"ContainerDied","Data":"2de211893f5be288b79b4a1fd765d0fa259d160fcb6f4cf42935dafe8605d5cd"} Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.763375 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-549bm"] Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.765048 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.767857 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.778785 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-549bm"] Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.852510 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-574gc\" (UniqueName: \"kubernetes.io/projected/8dd9b498-cece-42f6-88b7-65b998f421b0-kube-api-access-574gc\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.852580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-catalog-content\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.853019 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-utilities\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.956674 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-574gc\" (UniqueName: \"kubernetes.io/projected/8dd9b498-cece-42f6-88b7-65b998f421b0-kube-api-access-574gc\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.957283 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-catalog-content\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.957375 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-utilities\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.958952 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-utilities\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.960155 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-catalog-content\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.962780 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c9z79"] Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.973714 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.977768 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.984645 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9z79"] Oct 02 07:20:48 crc kubenswrapper[4960]: I1002 07:20:48.988009 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-574gc\" (UniqueName: \"kubernetes.io/projected/8dd9b498-cece-42f6-88b7-65b998f421b0-kube-api-access-574gc\") pod \"certified-operators-549bm\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.058368 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-catalog-content\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.058424 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-utilities\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.058487 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr57p\" (UniqueName: \"kubernetes.io/projected/7c4be745-c2cd-4a2b-844b-67572c49ca08-kube-api-access-tr57p\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.093682 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.159896 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr57p\" (UniqueName: \"kubernetes.io/projected/7c4be745-c2cd-4a2b-844b-67572c49ca08-kube-api-access-tr57p\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.160003 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-catalog-content\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.160044 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-utilities\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.160607 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-utilities\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.162812 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-catalog-content\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.184422 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr57p\" (UniqueName: \"kubernetes.io/projected/7c4be745-c2cd-4a2b-844b-67572c49ca08-kube-api-access-tr57p\") pod \"community-operators-c9z79\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.291628 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-549bm"] Oct 02 07:20:49 crc kubenswrapper[4960]: W1002 07:20:49.295313 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dd9b498_cece_42f6_88b7_65b998f421b0.slice/crio-6dcdc6bb5007d2a664eaeb5e63340b2bbedbe0970e71e3ff80e9e187b2485772 WatchSource:0}: Error finding container 6dcdc6bb5007d2a664eaeb5e63340b2bbedbe0970e71e3ff80e9e187b2485772: Status 404 returned error can't find the container with id 6dcdc6bb5007d2a664eaeb5e63340b2bbedbe0970e71e3ff80e9e187b2485772 Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.324030 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.622337 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9z79"] Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.622782 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw7n6" event={"ID":"95e601cc-5a5e-4630-9693-a7f6c42174d9","Type":"ContainerDied","Data":"bafc79410f5dc41ae394b51e799c7dde479a61398919f26d7f69a4605f38b73b"} Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.622351 4960 generic.go:334] "Generic (PLEG): container finished" podID="95e601cc-5a5e-4630-9693-a7f6c42174d9" containerID="bafc79410f5dc41ae394b51e799c7dde479a61398919f26d7f69a4605f38b73b" exitCode=0 Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.629631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq7lt" event={"ID":"ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7","Type":"ContainerStarted","Data":"e516bbcfc3f5d24c2f8bd5ba9c913e1f43364b3e1029fd7fcb5df9513539b26f"} Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.631084 4960 generic.go:334] "Generic (PLEG): container finished" podID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerID="be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71" exitCode=0 Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.631153 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-549bm" event={"ID":"8dd9b498-cece-42f6-88b7-65b998f421b0","Type":"ContainerDied","Data":"be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71"} Oct 02 07:20:49 crc kubenswrapper[4960]: I1002 07:20:49.631205 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-549bm" event={"ID":"8dd9b498-cece-42f6-88b7-65b998f421b0","Type":"ContainerStarted","Data":"6dcdc6bb5007d2a664eaeb5e63340b2bbedbe0970e71e3ff80e9e187b2485772"} Oct 02 07:20:49 crc kubenswrapper[4960]: W1002 07:20:49.631637 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c4be745_c2cd_4a2b_844b_67572c49ca08.slice/crio-6abcb5901e984ed810df9436b682ccd18f7f134e21014040e7e2ff9ccc2d6711 WatchSource:0}: Error finding container 6abcb5901e984ed810df9436b682ccd18f7f134e21014040e7e2ff9ccc2d6711: Status 404 returned error can't find the container with id 6abcb5901e984ed810df9436b682ccd18f7f134e21014040e7e2ff9ccc2d6711 Oct 02 07:20:50 crc kubenswrapper[4960]: I1002 07:20:50.351653 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rq7lt" podStartSLOduration=2.590268695 podStartE2EDuration="4.351636173s" podCreationTimestamp="2025-10-02 07:20:46 +0000 UTC" firstStartedPulling="2025-10-02 07:20:47.601563894 +0000 UTC m=+268.633510181" lastFinishedPulling="2025-10-02 07:20:49.362931372 +0000 UTC m=+270.394877659" observedRunningTime="2025-10-02 07:20:49.681889263 +0000 UTC m=+270.713835550" watchObservedRunningTime="2025-10-02 07:20:50.351636173 +0000 UTC m=+271.383582460" Oct 02 07:20:50 crc kubenswrapper[4960]: I1002 07:20:50.647522 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw7n6" event={"ID":"95e601cc-5a5e-4630-9693-a7f6c42174d9","Type":"ContainerStarted","Data":"764f06024e7e54cc3985579f33164dafeafac9001d82bd63266ed05a516fb8b4"} Oct 02 07:20:50 crc kubenswrapper[4960]: I1002 07:20:50.651742 4960 generic.go:334] "Generic (PLEG): container finished" podID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerID="f890a7397a96b6828e357e8a557dc2c9c20b1ed017cb37b36c03705f02910170" exitCode=0 Oct 02 07:20:50 crc kubenswrapper[4960]: I1002 07:20:50.651867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9z79" event={"ID":"7c4be745-c2cd-4a2b-844b-67572c49ca08","Type":"ContainerDied","Data":"f890a7397a96b6828e357e8a557dc2c9c20b1ed017cb37b36c03705f02910170"} Oct 02 07:20:50 crc kubenswrapper[4960]: I1002 07:20:50.651953 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9z79" event={"ID":"7c4be745-c2cd-4a2b-844b-67572c49ca08","Type":"ContainerStarted","Data":"6abcb5901e984ed810df9436b682ccd18f7f134e21014040e7e2ff9ccc2d6711"} Oct 02 07:20:50 crc kubenswrapper[4960]: I1002 07:20:50.678597 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nw7n6" podStartSLOduration=2.180084359 podStartE2EDuration="4.67856714s" podCreationTimestamp="2025-10-02 07:20:46 +0000 UTC" firstStartedPulling="2025-10-02 07:20:47.593107394 +0000 UTC m=+268.625053681" lastFinishedPulling="2025-10-02 07:20:50.091590175 +0000 UTC m=+271.123536462" observedRunningTime="2025-10-02 07:20:50.675712332 +0000 UTC m=+271.707658619" watchObservedRunningTime="2025-10-02 07:20:50.67856714 +0000 UTC m=+271.710513427" Oct 02 07:20:51 crc kubenswrapper[4960]: I1002 07:20:51.664202 4960 generic.go:334] "Generic (PLEG): container finished" podID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerID="80561f5224e962341e7ba53209b2d22283980c55363a175c43837b274185f8db" exitCode=0 Oct 02 07:20:51 crc kubenswrapper[4960]: I1002 07:20:51.664295 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9z79" event={"ID":"7c4be745-c2cd-4a2b-844b-67572c49ca08","Type":"ContainerDied","Data":"80561f5224e962341e7ba53209b2d22283980c55363a175c43837b274185f8db"} Oct 02 07:20:51 crc kubenswrapper[4960]: I1002 07:20:51.668223 4960 generic.go:334] "Generic (PLEG): container finished" podID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerID="a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb" exitCode=0 Oct 02 07:20:51 crc kubenswrapper[4960]: I1002 07:20:51.668436 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-549bm" event={"ID":"8dd9b498-cece-42f6-88b7-65b998f421b0","Type":"ContainerDied","Data":"a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb"} Oct 02 07:20:53 crc kubenswrapper[4960]: I1002 07:20:53.687796 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9z79" event={"ID":"7c4be745-c2cd-4a2b-844b-67572c49ca08","Type":"ContainerStarted","Data":"e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be"} Oct 02 07:20:53 crc kubenswrapper[4960]: I1002 07:20:53.690949 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-549bm" event={"ID":"8dd9b498-cece-42f6-88b7-65b998f421b0","Type":"ContainerStarted","Data":"544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267"} Oct 02 07:20:53 crc kubenswrapper[4960]: I1002 07:20:53.708857 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c9z79" podStartSLOduration=4.211947095 podStartE2EDuration="5.708802957s" podCreationTimestamp="2025-10-02 07:20:48 +0000 UTC" firstStartedPulling="2025-10-02 07:20:50.654340255 +0000 UTC m=+271.686286532" lastFinishedPulling="2025-10-02 07:20:52.151196107 +0000 UTC m=+273.183142394" observedRunningTime="2025-10-02 07:20:53.705371201 +0000 UTC m=+274.737317498" watchObservedRunningTime="2025-10-02 07:20:53.708802957 +0000 UTC m=+274.740749264" Oct 02 07:20:53 crc kubenswrapper[4960]: I1002 07:20:53.724526 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-549bm" podStartSLOduration=3.285127397 podStartE2EDuration="5.724497969s" podCreationTimestamp="2025-10-02 07:20:48 +0000 UTC" firstStartedPulling="2025-10-02 07:20:49.632747772 +0000 UTC m=+270.664694059" lastFinishedPulling="2025-10-02 07:20:52.072118304 +0000 UTC m=+273.104064631" observedRunningTime="2025-10-02 07:20:53.72158112 +0000 UTC m=+274.753527407" watchObservedRunningTime="2025-10-02 07:20:53.724497969 +0000 UTC m=+274.756444266" Oct 02 07:20:56 crc kubenswrapper[4960]: I1002 07:20:56.686178 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:56 crc kubenswrapper[4960]: I1002 07:20:56.686873 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:56 crc kubenswrapper[4960]: I1002 07:20:56.759750 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:56 crc kubenswrapper[4960]: I1002 07:20:56.805450 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rq7lt" Oct 02 07:20:56 crc kubenswrapper[4960]: I1002 07:20:56.899370 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:56 crc kubenswrapper[4960]: I1002 07:20:56.899773 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:56 crc kubenswrapper[4960]: I1002 07:20:56.954230 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:57 crc kubenswrapper[4960]: I1002 07:20:57.758557 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nw7n6" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.095290 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.095357 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.133417 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.324741 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.324836 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.377893 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.794398 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:20:59 crc kubenswrapper[4960]: I1002 07:20:59.870246 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:22:29 crc kubenswrapper[4960]: I1002 07:22:29.150212 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:22:29 crc kubenswrapper[4960]: I1002 07:22:29.151383 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:22:59 crc kubenswrapper[4960]: I1002 07:22:59.150003 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:22:59 crc kubenswrapper[4960]: I1002 07:22:59.150901 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.053774 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n2kkg"] Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.055212 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.070579 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n2kkg"] Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184439 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-registry-certificates\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184516 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184567 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184600 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgk7g\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-kube-api-access-vgk7g\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184626 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184776 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-trusted-ca\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184802 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-registry-tls\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.184821 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-bound-sa-token\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.208123 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.286292 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.286814 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgk7g\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-kube-api-access-vgk7g\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.286840 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.286882 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-trusted-ca\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.286899 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-registry-tls\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.286921 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-bound-sa-token\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.286962 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-registry-certificates\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.288364 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-registry-certificates\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.288783 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.290271 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-trusted-ca\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.297052 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.297181 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-registry-tls\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.308411 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-bound-sa-token\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.309135 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgk7g\" (UniqueName: \"kubernetes.io/projected/1196d962-d0f4-42d6-89fe-bd18f9bc59ef-kube-api-access-vgk7g\") pod \"image-registry-66df7c8f76-n2kkg\" (UID: \"1196d962-d0f4-42d6-89fe-bd18f9bc59ef\") " pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.380188 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.589965 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n2kkg"] Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.827270 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" event={"ID":"1196d962-d0f4-42d6-89fe-bd18f9bc59ef","Type":"ContainerStarted","Data":"ba451b390f9c624ae6faba2c3376a2e84e5bbcbd413c58b4a78fcf0d7ded857f"} Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.827655 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" event={"ID":"1196d962-d0f4-42d6-89fe-bd18f9bc59ef","Type":"ContainerStarted","Data":"dbcf623b421d1051e38cbf1a5b455d5bd5518afff83cbd5057d2b3090a9ad6a8"} Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.827679 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:28 crc kubenswrapper[4960]: I1002 07:23:28.847696 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" podStartSLOduration=0.847673783 podStartE2EDuration="847.673783ms" podCreationTimestamp="2025-10-02 07:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:23:28.847061164 +0000 UTC m=+429.879007461" watchObservedRunningTime="2025-10-02 07:23:28.847673783 +0000 UTC m=+429.879620070" Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.150590 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.150686 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.150768 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.151718 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e05a2b35123183dd07d94c3c5aa895e60ba41f5e31b710be48e94c02186f6ea1"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.151822 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://e05a2b35123183dd07d94c3c5aa895e60ba41f5e31b710be48e94c02186f6ea1" gracePeriod=600 Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.835774 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="e05a2b35123183dd07d94c3c5aa895e60ba41f5e31b710be48e94c02186f6ea1" exitCode=0 Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.835862 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"e05a2b35123183dd07d94c3c5aa895e60ba41f5e31b710be48e94c02186f6ea1"} Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.836307 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"078c8db152d306f847e249781287a9e53d5a972cc64796b1689df8c3d5c083f5"} Oct 02 07:23:29 crc kubenswrapper[4960]: I1002 07:23:29.836330 4960 scope.go:117] "RemoveContainer" containerID="a261499b8a91874674a3bf92c71e5619828221f58b6d25432cc4414cedf46f25" Oct 02 07:23:48 crc kubenswrapper[4960]: I1002 07:23:48.388102 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-n2kkg" Oct 02 07:23:48 crc kubenswrapper[4960]: I1002 07:23:48.449831 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8md6w"] Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.524597 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" podUID="a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" containerName="registry" containerID="cri-o://b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e" gracePeriod=30 Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.910715 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.974605 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-certificates\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.974698 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-trusted-ca\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.974775 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-ca-trust-extracted\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.975163 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.975208 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-tls\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.975320 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hqlf\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-kube-api-access-2hqlf\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.975388 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-installation-pull-secrets\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.975434 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-bound-sa-token\") pod \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\" (UID: \"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e\") " Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.975831 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.976581 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.984048 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.984699 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.987604 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.989520 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-kube-api-access-2hqlf" (OuterVolumeSpecName: "kube-api-access-2hqlf") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "kube-api-access-2hqlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:24:13 crc kubenswrapper[4960]: I1002 07:24:13.994734 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.017016 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" (UID: "a9c46af1-aa1f-46cc-88b5-5ac0eb50337e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.077361 4960 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.077413 4960 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.077436 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hqlf\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-kube-api-access-2hqlf\") on node \"crc\" DevicePath \"\"" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.077460 4960 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.077478 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.077495 4960 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.077512 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.149563 4960 generic.go:334] "Generic (PLEG): container finished" podID="a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" containerID="b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e" exitCode=0 Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.149643 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.149687 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" event={"ID":"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e","Type":"ContainerDied","Data":"b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e"} Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.150110 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8md6w" event={"ID":"a9c46af1-aa1f-46cc-88b5-5ac0eb50337e","Type":"ContainerDied","Data":"ba1f051df8b4c74494d985b69e4f5da01a2d89f98e38a60a7b582ae878534ecb"} Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.150194 4960 scope.go:117] "RemoveContainer" containerID="b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.175197 4960 scope.go:117] "RemoveContainer" containerID="b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e" Oct 02 07:24:14 crc kubenswrapper[4960]: E1002 07:24:14.176500 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e\": container with ID starting with b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e not found: ID does not exist" containerID="b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.176565 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e"} err="failed to get container status \"b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e\": rpc error: code = NotFound desc = could not find container \"b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e\": container with ID starting with b3fe20d37a0726e3892dd491881279c163a44004a52bb41908a5747009760c3e not found: ID does not exist" Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.199698 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8md6w"] Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.206037 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8md6w"] Oct 02 07:24:14 crc kubenswrapper[4960]: I1002 07:24:14.339612 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" path="/var/lib/kubelet/pods/a9c46af1-aa1f-46cc-88b5-5ac0eb50337e/volumes" Oct 02 07:25:29 crc kubenswrapper[4960]: I1002 07:25:29.150054 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:25:29 crc kubenswrapper[4960]: I1002 07:25:29.151027 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:25:59 crc kubenswrapper[4960]: I1002 07:25:59.150443 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:25:59 crc kubenswrapper[4960]: I1002 07:25:59.151525 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.810182 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-74597"] Oct 02 07:26:10 crc kubenswrapper[4960]: E1002 07:26:10.812266 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" containerName="registry" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.812335 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" containerName="registry" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.812482 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9c46af1-aa1f-46cc-88b5-5ac0eb50337e" containerName="registry" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.813019 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.815136 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.815537 4960 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cp255" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.815758 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.819375 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vn5rf"] Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.820301 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vn5rf" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.823182 4960 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-2kt4c" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.833635 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-74597"] Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.837398 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vn5rf"] Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.847563 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zh4pv"] Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.849105 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.851029 4960 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-pl7lq" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.863595 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zh4pv"] Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.995140 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkqsd\" (UniqueName: \"kubernetes.io/projected/c9ca51b7-0180-491c-9d6d-730bea696f09-kube-api-access-lkqsd\") pod \"cert-manager-cainjector-7f985d654d-74597\" (UID: \"c9ca51b7-0180-491c-9d6d-730bea696f09\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.995195 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvfnt\" (UniqueName: \"kubernetes.io/projected/3a34826f-fbd3-4abe-a67f-adc680605fb9-kube-api-access-nvfnt\") pod \"cert-manager-webhook-5655c58dd6-zh4pv\" (UID: \"3a34826f-fbd3-4abe-a67f-adc680605fb9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" Oct 02 07:26:10 crc kubenswrapper[4960]: I1002 07:26:10.995236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79xbt\" (UniqueName: \"kubernetes.io/projected/aa123932-baa6-435e-b311-e6a62417f39f-kube-api-access-79xbt\") pod \"cert-manager-5b446d88c5-vn5rf\" (UID: \"aa123932-baa6-435e-b311-e6a62417f39f\") " pod="cert-manager/cert-manager-5b446d88c5-vn5rf" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.096863 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkqsd\" (UniqueName: \"kubernetes.io/projected/c9ca51b7-0180-491c-9d6d-730bea696f09-kube-api-access-lkqsd\") pod \"cert-manager-cainjector-7f985d654d-74597\" (UID: \"c9ca51b7-0180-491c-9d6d-730bea696f09\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.097496 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvfnt\" (UniqueName: \"kubernetes.io/projected/3a34826f-fbd3-4abe-a67f-adc680605fb9-kube-api-access-nvfnt\") pod \"cert-manager-webhook-5655c58dd6-zh4pv\" (UID: \"3a34826f-fbd3-4abe-a67f-adc680605fb9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.097559 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79xbt\" (UniqueName: \"kubernetes.io/projected/aa123932-baa6-435e-b311-e6a62417f39f-kube-api-access-79xbt\") pod \"cert-manager-5b446d88c5-vn5rf\" (UID: \"aa123932-baa6-435e-b311-e6a62417f39f\") " pod="cert-manager/cert-manager-5b446d88c5-vn5rf" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.118736 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkqsd\" (UniqueName: \"kubernetes.io/projected/c9ca51b7-0180-491c-9d6d-730bea696f09-kube-api-access-lkqsd\") pod \"cert-manager-cainjector-7f985d654d-74597\" (UID: \"c9ca51b7-0180-491c-9d6d-730bea696f09\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.118750 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvfnt\" (UniqueName: \"kubernetes.io/projected/3a34826f-fbd3-4abe-a67f-adc680605fb9-kube-api-access-nvfnt\") pod \"cert-manager-webhook-5655c58dd6-zh4pv\" (UID: \"3a34826f-fbd3-4abe-a67f-adc680605fb9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.119965 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79xbt\" (UniqueName: \"kubernetes.io/projected/aa123932-baa6-435e-b311-e6a62417f39f-kube-api-access-79xbt\") pod \"cert-manager-5b446d88c5-vn5rf\" (UID: \"aa123932-baa6-435e-b311-e6a62417f39f\") " pod="cert-manager/cert-manager-5b446d88c5-vn5rf" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.135905 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.145728 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vn5rf" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.165415 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.657295 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-74597"] Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.667550 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vn5rf"] Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.673396 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.688281 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zh4pv"] Oct 02 07:26:11 crc kubenswrapper[4960]: W1002 07:26:11.694720 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a34826f_fbd3_4abe_a67f_adc680605fb9.slice/crio-2f25274441719c4ee51873103f22ebcf6db8afe2560763215a8dbae878a20a74 WatchSource:0}: Error finding container 2f25274441719c4ee51873103f22ebcf6db8afe2560763215a8dbae878a20a74: Status 404 returned error can't find the container with id 2f25274441719c4ee51873103f22ebcf6db8afe2560763215a8dbae878a20a74 Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.973746 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vn5rf" event={"ID":"aa123932-baa6-435e-b311-e6a62417f39f","Type":"ContainerStarted","Data":"2871db887ab9f1855522597f86de769b6935fcf872a2ce280eb32faa8c50103a"} Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.975568 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" event={"ID":"c9ca51b7-0180-491c-9d6d-730bea696f09","Type":"ContainerStarted","Data":"15c3204bd4bff833565937784c655cfcc16cf60e2d7c5b1499e85a3f45e4c37b"} Oct 02 07:26:11 crc kubenswrapper[4960]: I1002 07:26:11.977047 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" event={"ID":"3a34826f-fbd3-4abe-a67f-adc680605fb9","Type":"ContainerStarted","Data":"2f25274441719c4ee51873103f22ebcf6db8afe2560763215a8dbae878a20a74"} Oct 02 07:26:16 crc kubenswrapper[4960]: I1002 07:26:16.003670 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vn5rf" event={"ID":"aa123932-baa6-435e-b311-e6a62417f39f","Type":"ContainerStarted","Data":"cb9216216061516647fdca2a5a48f2ce6284b8aea1829360d1f594a298e451d6"} Oct 02 07:26:16 crc kubenswrapper[4960]: I1002 07:26:16.006255 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" event={"ID":"c9ca51b7-0180-491c-9d6d-730bea696f09","Type":"ContainerStarted","Data":"f64922a97f1598423754813411beeb2c395f5abe7c2a780af63fb24670c0070d"} Oct 02 07:26:16 crc kubenswrapper[4960]: I1002 07:26:16.008335 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" event={"ID":"3a34826f-fbd3-4abe-a67f-adc680605fb9","Type":"ContainerStarted","Data":"9aeec69ef50b2dd27f096059a46465acaceab9c2e8d7b5f68c60957ac981f85c"} Oct 02 07:26:16 crc kubenswrapper[4960]: I1002 07:26:16.008468 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" Oct 02 07:26:16 crc kubenswrapper[4960]: I1002 07:26:16.027667 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-vn5rf" podStartSLOduration=2.432805473 podStartE2EDuration="6.027642169s" podCreationTimestamp="2025-10-02 07:26:10 +0000 UTC" firstStartedPulling="2025-10-02 07:26:11.681480247 +0000 UTC m=+592.713426534" lastFinishedPulling="2025-10-02 07:26:15.276316913 +0000 UTC m=+596.308263230" observedRunningTime="2025-10-02 07:26:16.023558838 +0000 UTC m=+597.055505135" watchObservedRunningTime="2025-10-02 07:26:16.027642169 +0000 UTC m=+597.059588476" Oct 02 07:26:16 crc kubenswrapper[4960]: I1002 07:26:16.044879 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" podStartSLOduration=2.482035991 podStartE2EDuration="6.044856687s" podCreationTimestamp="2025-10-02 07:26:10 +0000 UTC" firstStartedPulling="2025-10-02 07:26:11.698051317 +0000 UTC m=+592.729997614" lastFinishedPulling="2025-10-02 07:26:15.260871993 +0000 UTC m=+596.292818310" observedRunningTime="2025-10-02 07:26:16.042141843 +0000 UTC m=+597.074088130" watchObservedRunningTime="2025-10-02 07:26:16.044856687 +0000 UTC m=+597.076802974" Oct 02 07:26:16 crc kubenswrapper[4960]: I1002 07:26:16.063350 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-74597" podStartSLOduration=2.465579464 podStartE2EDuration="6.063322519s" podCreationTimestamp="2025-10-02 07:26:10 +0000 UTC" firstStartedPulling="2025-10-02 07:26:11.673077648 +0000 UTC m=+592.705023935" lastFinishedPulling="2025-10-02 07:26:15.270820693 +0000 UTC m=+596.302766990" observedRunningTime="2025-10-02 07:26:16.059951987 +0000 UTC m=+597.091898304" watchObservedRunningTime="2025-10-02 07:26:16.063322519 +0000 UTC m=+597.095268846" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.170870 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-zh4pv" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.341189 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8dmqg"] Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.342226 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-controller" containerID="cri-o://af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.342717 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="northd" containerID="cri-o://92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.342904 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.343045 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-node" containerID="cri-o://a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.343047 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="sbdb" containerID="cri-o://979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.343124 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="nbdb" containerID="cri-o://546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.343152 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-acl-logging" containerID="cri-o://ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.390267 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" containerID="cri-o://3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" gracePeriod=30 Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.691919 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/3.log" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.694189 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovn-acl-logging/0.log" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.694774 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovn-controller/0.log" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.695320 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.755838 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-h795z"] Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.756389 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.756467 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.756530 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-node" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.756583 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-node" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.756637 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="nbdb" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.756686 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="nbdb" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.756825 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="sbdb" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.756877 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="sbdb" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.756944 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757136 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.757196 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-acl-logging" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757242 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-acl-logging" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.757289 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kubecfg-setup" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757338 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kubecfg-setup" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.757383 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757428 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.757486 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757538 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.757583 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757636 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.757692 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="northd" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757741 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="northd" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.757799 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.757845 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758011 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758079 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758143 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="sbdb" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758229 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758283 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="northd" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758368 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758418 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovn-acl-logging" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758475 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758539 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758591 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="nbdb" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758643 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="kube-rbac-proxy-node" Oct 02 07:26:21 crc kubenswrapper[4960]: E1002 07:26:21.758803 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.758860 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.759062 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerName="ovnkube-controller" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.760816 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.863426 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-systemd\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.863761 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-netns\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.863862 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-ovn-kubernetes\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864055 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-kubelet\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864165 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-config\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864258 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-openvswitch\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864359 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-systemd-units\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864470 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-ovn\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864597 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-slash\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-netd\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864851 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.863944 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864014 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864190 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864337 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864434 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864561 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864676 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-slash" (OuterVolumeSpecName: "host-slash") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864815 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864863 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864923 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.864945 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ab5426-0159-4610-b378-f4c2e20b82b9-ovn-node-metrics-cert\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865443 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-log-socket\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865516 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-bin\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865539 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-log-socket" (OuterVolumeSpecName: "log-socket") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865558 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865591 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-env-overrides\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865703 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-node-log\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865752 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-etc-openvswitch\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865781 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-node-log" (OuterVolumeSpecName: "node-log") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.865842 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.866069 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxd2c\" (UniqueName: \"kubernetes.io/projected/41ab5426-0159-4610-b378-f4c2e20b82b9-kube-api-access-dxd2c\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.866162 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-var-lib-openvswitch\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.866252 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.866333 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-script-lib\") pod \"41ab5426-0159-4610-b378-f4c2e20b82b9\" (UID: \"41ab5426-0159-4610-b378-f4c2e20b82b9\") " Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.866534 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.866906 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867293 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovn-node-metrics-cert\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867348 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-kubelet\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867380 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-env-overrides\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867509 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-run-ovn-kubernetes\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867619 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-cni-netd\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-ovn\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867747 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.867825 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868424 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-var-lib-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868554 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovnkube-config\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868615 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-slash\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868698 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-log-socket\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868735 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovnkube-script-lib\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868799 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-cni-bin\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868832 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flkg9\" (UniqueName: \"kubernetes.io/projected/0741294a-e888-44b7-8d7c-0b5162d72ca9-kube-api-access-flkg9\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.868958 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-run-netns\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869062 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-systemd-units\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869156 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-etc-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869333 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-systemd\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869434 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-node-log\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869677 4960 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869716 4960 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869736 4960 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869757 4960 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869774 4960 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869793 4960 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869810 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869827 4960 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869845 4960 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869861 4960 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869878 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ab5426-0159-4610-b378-f4c2e20b82b9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869895 4960 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869913 4960 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869938 4960 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869955 4960 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.869972 4960 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.870013 4960 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.871048 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ab5426-0159-4610-b378-f4c2e20b82b9-kube-api-access-dxd2c" (OuterVolumeSpecName: "kube-api-access-dxd2c") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "kube-api-access-dxd2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.871289 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ab5426-0159-4610-b378-f4c2e20b82b9-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.877827 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "41ab5426-0159-4610-b378-f4c2e20b82b9" (UID: "41ab5426-0159-4610-b378-f4c2e20b82b9"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.970699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-var-lib-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971070 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovnkube-config\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971188 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-slash\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971276 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-log-socket\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971341 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-log-socket\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.970803 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-var-lib-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971266 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-slash\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971356 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovnkube-script-lib\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971550 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-cni-bin\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971621 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flkg9\" (UniqueName: \"kubernetes.io/projected/0741294a-e888-44b7-8d7c-0b5162d72ca9-kube-api-access-flkg9\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971705 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-run-netns\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971800 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-systemd-units\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971861 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-systemd-units\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971643 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-cni-bin\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971873 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-etc-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.971826 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-run-netns\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972021 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovnkube-config\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972051 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-systemd\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972079 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-systemd\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972114 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-node-log\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972142 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovnkube-script-lib\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972187 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-node-log\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972192 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovn-node-metrics-cert\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972310 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-kubelet\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972327 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-env-overrides\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972372 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-kubelet\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972408 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-run-ovn-kubernetes\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972449 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-cni-netd\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972470 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-ovn\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972500 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972522 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972594 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ab5426-0159-4610-b378-f4c2e20b82b9-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972607 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxd2c\" (UniqueName: \"kubernetes.io/projected/41ab5426-0159-4610-b378-f4c2e20b82b9-kube-api-access-dxd2c\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972620 4960 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ab5426-0159-4610-b378-f4c2e20b82b9-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972660 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972686 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-run-ovn-kubernetes\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.972707 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-host-cni-netd\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.973051 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-etc-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.973067 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0741294a-e888-44b7-8d7c-0b5162d72ca9-env-overrides\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.973111 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-ovn\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.973143 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0741294a-e888-44b7-8d7c-0b5162d72ca9-run-openvswitch\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.975952 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0741294a-e888-44b7-8d7c-0b5162d72ca9-ovn-node-metrics-cert\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:21 crc kubenswrapper[4960]: I1002 07:26:21.999787 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flkg9\" (UniqueName: \"kubernetes.io/projected/0741294a-e888-44b7-8d7c-0b5162d72ca9-kube-api-access-flkg9\") pod \"ovnkube-node-h795z\" (UID: \"0741294a-e888-44b7-8d7c-0b5162d72ca9\") " pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.066113 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovnkube-controller/3.log" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.069123 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovn-acl-logging/0.log" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.069634 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8dmqg_41ab5426-0159-4610-b378-f4c2e20b82b9/ovn-controller/0.log" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072296 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" exitCode=0 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072360 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" exitCode=0 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072377 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" exitCode=0 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072375 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072410 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" exitCode=0 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072428 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" exitCode=0 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072448 4960 scope.go:117] "RemoveContainer" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072449 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" exitCode=0 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072556 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" exitCode=143 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072431 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072575 4960 generic.go:334] "Generic (PLEG): container finished" podID="41ab5426-0159-4610-b378-f4c2e20b82b9" containerID="af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" exitCode=143 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072590 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072606 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072619 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072646 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072659 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072666 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072673 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072680 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072686 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072692 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072699 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072705 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072712 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072722 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072728 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072734 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072739 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072745 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072751 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072757 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072763 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072770 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072776 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072783 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072793 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072800 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072805 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072813 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072818 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072823 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072829 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072834 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072385 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072917 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072839 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072960 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8dmqg" event={"ID":"41ab5426-0159-4610-b378-f4c2e20b82b9","Type":"ContainerDied","Data":"2fe94077b19b160088ae5787181c9eb342d2e0beea62869eb4a1a253aeace771"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.072995 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073002 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073007 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073012 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073018 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073023 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073028 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073034 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073050 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.073055 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.075358 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/2.log" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.077239 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/1.log" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.077311 4960 generic.go:334] "Generic (PLEG): container finished" podID="c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e" containerID="58bee76173a69e8b6bf08b9a1abd3626ec971618f989710817a49de54f5243e2" exitCode=2 Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.077379 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerDied","Data":"58bee76173a69e8b6bf08b9a1abd3626ec971618f989710817a49de54f5243e2"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.077430 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3"} Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.078620 4960 scope.go:117] "RemoveContainer" containerID="58bee76173a69e8b6bf08b9a1abd3626ec971618f989710817a49de54f5243e2" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.079094 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2dqxx_openshift-multus(c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e)\"" pod="openshift-multus/multus-2dqxx" podUID="c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.093553 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:26:22 crc kubenswrapper[4960]: W1002 07:26:22.103409 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0741294a_e888_44b7_8d7c_0b5162d72ca9.slice/crio-8da7673179caa9d3e5e5bf88570fe3bcbef23a296daf8201afba1fcc0544b36e WatchSource:0}: Error finding container 8da7673179caa9d3e5e5bf88570fe3bcbef23a296daf8201afba1fcc0544b36e: Status 404 returned error can't find the container with id 8da7673179caa9d3e5e5bf88570fe3bcbef23a296daf8201afba1fcc0544b36e Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.123509 4960 scope.go:117] "RemoveContainer" containerID="979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.123722 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8dmqg"] Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.139235 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8dmqg"] Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.159719 4960 scope.go:117] "RemoveContainer" containerID="546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.185200 4960 scope.go:117] "RemoveContainer" containerID="92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.220593 4960 scope.go:117] "RemoveContainer" containerID="884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.233555 4960 scope.go:117] "RemoveContainer" containerID="a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.252067 4960 scope.go:117] "RemoveContainer" containerID="ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.271535 4960 scope.go:117] "RemoveContainer" containerID="af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.340090 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41ab5426-0159-4610-b378-f4c2e20b82b9" path="/var/lib/kubelet/pods/41ab5426-0159-4610-b378-f4c2e20b82b9/volumes" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.346178 4960 scope.go:117] "RemoveContainer" containerID="67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.366181 4960 scope.go:117] "RemoveContainer" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.366588 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": container with ID starting with 3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d not found: ID does not exist" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.366618 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} err="failed to get container status \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": rpc error: code = NotFound desc = could not find container \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": container with ID starting with 3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.366642 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.368187 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": container with ID starting with 630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f not found: ID does not exist" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.368413 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} err="failed to get container status \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": rpc error: code = NotFound desc = could not find container \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": container with ID starting with 630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.368502 4960 scope.go:117] "RemoveContainer" containerID="979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.368991 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": container with ID starting with 979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517 not found: ID does not exist" containerID="979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.369103 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} err="failed to get container status \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": rpc error: code = NotFound desc = could not find container \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": container with ID starting with 979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.369205 4960 scope.go:117] "RemoveContainer" containerID="546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.369577 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": container with ID starting with 546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f not found: ID does not exist" containerID="546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.369625 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} err="failed to get container status \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": rpc error: code = NotFound desc = could not find container \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": container with ID starting with 546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.369656 4960 scope.go:117] "RemoveContainer" containerID="92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.369960 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": container with ID starting with 92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628 not found: ID does not exist" containerID="92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.370078 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} err="failed to get container status \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": rpc error: code = NotFound desc = could not find container \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": container with ID starting with 92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.370155 4960 scope.go:117] "RemoveContainer" containerID="884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.370431 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": container with ID starting with 884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe not found: ID does not exist" containerID="884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.370455 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} err="failed to get container status \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": rpc error: code = NotFound desc = could not find container \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": container with ID starting with 884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.370470 4960 scope.go:117] "RemoveContainer" containerID="a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.370817 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": container with ID starting with a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e not found: ID does not exist" containerID="a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.370848 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} err="failed to get container status \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": rpc error: code = NotFound desc = could not find container \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": container with ID starting with a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.370863 4960 scope.go:117] "RemoveContainer" containerID="ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.371600 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": container with ID starting with ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4 not found: ID does not exist" containerID="ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.371654 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} err="failed to get container status \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": rpc error: code = NotFound desc = could not find container \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": container with ID starting with ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.371685 4960 scope.go:117] "RemoveContainer" containerID="af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.372033 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": container with ID starting with af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8 not found: ID does not exist" containerID="af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.372061 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} err="failed to get container status \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": rpc error: code = NotFound desc = could not find container \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": container with ID starting with af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.372080 4960 scope.go:117] "RemoveContainer" containerID="67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741" Oct 02 07:26:22 crc kubenswrapper[4960]: E1002 07:26:22.374215 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": container with ID starting with 67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741 not found: ID does not exist" containerID="67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.374265 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} err="failed to get container status \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": rpc error: code = NotFound desc = could not find container \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": container with ID starting with 67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.374296 4960 scope.go:117] "RemoveContainer" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.374892 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} err="failed to get container status \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": rpc error: code = NotFound desc = could not find container \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": container with ID starting with 3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.374917 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.375155 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} err="failed to get container status \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": rpc error: code = NotFound desc = could not find container \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": container with ID starting with 630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.375175 4960 scope.go:117] "RemoveContainer" containerID="979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.375553 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} err="failed to get container status \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": rpc error: code = NotFound desc = could not find container \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": container with ID starting with 979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.375572 4960 scope.go:117] "RemoveContainer" containerID="546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.378921 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} err="failed to get container status \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": rpc error: code = NotFound desc = could not find container \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": container with ID starting with 546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.378942 4960 scope.go:117] "RemoveContainer" containerID="92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.379360 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} err="failed to get container status \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": rpc error: code = NotFound desc = could not find container \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": container with ID starting with 92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.379388 4960 scope.go:117] "RemoveContainer" containerID="884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.379738 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} err="failed to get container status \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": rpc error: code = NotFound desc = could not find container \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": container with ID starting with 884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.379836 4960 scope.go:117] "RemoveContainer" containerID="a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.380178 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} err="failed to get container status \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": rpc error: code = NotFound desc = could not find container \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": container with ID starting with a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.380290 4960 scope.go:117] "RemoveContainer" containerID="ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.380606 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} err="failed to get container status \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": rpc error: code = NotFound desc = could not find container \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": container with ID starting with ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.380624 4960 scope.go:117] "RemoveContainer" containerID="af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.380865 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} err="failed to get container status \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": rpc error: code = NotFound desc = could not find container \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": container with ID starting with af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.380962 4960 scope.go:117] "RemoveContainer" containerID="67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.381282 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} err="failed to get container status \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": rpc error: code = NotFound desc = could not find container \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": container with ID starting with 67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.381392 4960 scope.go:117] "RemoveContainer" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.381689 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} err="failed to get container status \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": rpc error: code = NotFound desc = could not find container \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": container with ID starting with 3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.381784 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.382119 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} err="failed to get container status \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": rpc error: code = NotFound desc = could not find container \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": container with ID starting with 630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.382247 4960 scope.go:117] "RemoveContainer" containerID="979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.382527 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} err="failed to get container status \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": rpc error: code = NotFound desc = could not find container \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": container with ID starting with 979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.382548 4960 scope.go:117] "RemoveContainer" containerID="546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.382849 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} err="failed to get container status \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": rpc error: code = NotFound desc = could not find container \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": container with ID starting with 546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.382897 4960 scope.go:117] "RemoveContainer" containerID="92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.383204 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} err="failed to get container status \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": rpc error: code = NotFound desc = could not find container \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": container with ID starting with 92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.383227 4960 scope.go:117] "RemoveContainer" containerID="884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.383488 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} err="failed to get container status \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": rpc error: code = NotFound desc = could not find container \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": container with ID starting with 884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.383536 4960 scope.go:117] "RemoveContainer" containerID="a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.383829 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} err="failed to get container status \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": rpc error: code = NotFound desc = could not find container \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": container with ID starting with a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.383916 4960 scope.go:117] "RemoveContainer" containerID="ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.384278 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} err="failed to get container status \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": rpc error: code = NotFound desc = could not find container \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": container with ID starting with ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.384353 4960 scope.go:117] "RemoveContainer" containerID="af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.384839 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} err="failed to get container status \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": rpc error: code = NotFound desc = could not find container \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": container with ID starting with af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.385048 4960 scope.go:117] "RemoveContainer" containerID="67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.385296 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} err="failed to get container status \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": rpc error: code = NotFound desc = could not find container \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": container with ID starting with 67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.385379 4960 scope.go:117] "RemoveContainer" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.385676 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} err="failed to get container status \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": rpc error: code = NotFound desc = could not find container \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": container with ID starting with 3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.385788 4960 scope.go:117] "RemoveContainer" containerID="630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.386201 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f"} err="failed to get container status \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": rpc error: code = NotFound desc = could not find container \"630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f\": container with ID starting with 630fb2a9e82f2d9cb2a9f6699f2825c822621bb66821cd703e722d6397e1348f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.386240 4960 scope.go:117] "RemoveContainer" containerID="979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.386493 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517"} err="failed to get container status \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": rpc error: code = NotFound desc = could not find container \"979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517\": container with ID starting with 979491403c5f7fae305037be35b8482553f102853197ad7a8aab28a1ab512517 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.386574 4960 scope.go:117] "RemoveContainer" containerID="546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.386827 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f"} err="failed to get container status \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": rpc error: code = NotFound desc = could not find container \"546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f\": container with ID starting with 546a0be5f94223c36ec68b22681aea6ef737718a0f857c2283c01334c811240f not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.386849 4960 scope.go:117] "RemoveContainer" containerID="92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.387097 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628"} err="failed to get container status \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": rpc error: code = NotFound desc = could not find container \"92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628\": container with ID starting with 92c9edf97f6a903a9fc0d7db830a160992dac16b2f4b1536f6c6cfaae5cbe628 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.387186 4960 scope.go:117] "RemoveContainer" containerID="884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.387485 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe"} err="failed to get container status \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": rpc error: code = NotFound desc = could not find container \"884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe\": container with ID starting with 884ef8a253869ef2c5f0ecdff893e4acaa9de1df526b0bf88708591dce1baebe not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.387501 4960 scope.go:117] "RemoveContainer" containerID="a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.387761 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e"} err="failed to get container status \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": rpc error: code = NotFound desc = could not find container \"a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e\": container with ID starting with a8b08de18c03d1be75603ba20c1385a2fb32af09ca84163d83bd99b9a83c024e not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.387837 4960 scope.go:117] "RemoveContainer" containerID="ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.388196 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4"} err="failed to get container status \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": rpc error: code = NotFound desc = could not find container \"ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4\": container with ID starting with ffd0d4b5d987fb74c387d217c3042ba9e329c6b50d30aea1a4ccc935132e69a4 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.388237 4960 scope.go:117] "RemoveContainer" containerID="af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.388569 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8"} err="failed to get container status \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": rpc error: code = NotFound desc = could not find container \"af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8\": container with ID starting with af9ef77c219922d9092c35c7169f6dc3f36b72fa49b9dc26f3476341df2fb1e8 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.388648 4960 scope.go:117] "RemoveContainer" containerID="67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.389033 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741"} err="failed to get container status \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": rpc error: code = NotFound desc = could not find container \"67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741\": container with ID starting with 67ed993fd09e4c173aa135c864cc0aff9599ab21a51a3dbc0b68f66f0668c741 not found: ID does not exist" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.389069 4960 scope.go:117] "RemoveContainer" containerID="3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d" Oct 02 07:26:22 crc kubenswrapper[4960]: I1002 07:26:22.389372 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d"} err="failed to get container status \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": rpc error: code = NotFound desc = could not find container \"3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d\": container with ID starting with 3eeb7029f8620408dc3656373ed33173ac8177324fd45c2b97d48d72ed8dc24d not found: ID does not exist" Oct 02 07:26:23 crc kubenswrapper[4960]: I1002 07:26:23.087924 4960 generic.go:334] "Generic (PLEG): container finished" podID="0741294a-e888-44b7-8d7c-0b5162d72ca9" containerID="148dc988023e362ad20695dc9d198f99f91619050602c5f1d903ab0503c51b10" exitCode=0 Oct 02 07:26:23 crc kubenswrapper[4960]: I1002 07:26:23.087970 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerDied","Data":"148dc988023e362ad20695dc9d198f99f91619050602c5f1d903ab0503c51b10"} Oct 02 07:26:23 crc kubenswrapper[4960]: I1002 07:26:23.088010 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"8da7673179caa9d3e5e5bf88570fe3bcbef23a296daf8201afba1fcc0544b36e"} Oct 02 07:26:24 crc kubenswrapper[4960]: I1002 07:26:24.099362 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"b3e8fc9fe27e34016f5b46bf22514685a9737e69406bd29c4f2d9b8201d95e2a"} Oct 02 07:26:24 crc kubenswrapper[4960]: I1002 07:26:24.101331 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"99f02c90eccfddaff7e38c272295a5d911c23c4acc9d859b8ad4772c5289c521"} Oct 02 07:26:24 crc kubenswrapper[4960]: I1002 07:26:24.101374 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"3a89ab239f06f62ee709d9c8b501f2ea2a7ae0de7d4e67ebb84f7ec041bf4981"} Oct 02 07:26:24 crc kubenswrapper[4960]: I1002 07:26:24.101412 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"f5f51deec6cdd5eb7052a5283bd5e0fbd48d6c0344f45782e750381b7b2c0a6b"} Oct 02 07:26:24 crc kubenswrapper[4960]: I1002 07:26:24.101435 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"b443dc5b2e4d8bf7708c5fe23d8fde321f157d43ff12793d5961d8be973c0be7"} Oct 02 07:26:24 crc kubenswrapper[4960]: I1002 07:26:24.101453 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"360ffdf8d2258d505324750e6b769146461a65de4bd18c72155bfec8740bc2f6"} Oct 02 07:26:27 crc kubenswrapper[4960]: I1002 07:26:27.122171 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"e0ce2d3460e09055652e60e4fd713e9b1aef349776e1ac0fa39bfc91beee0386"} Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.137812 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" event={"ID":"0741294a-e888-44b7-8d7c-0b5162d72ca9","Type":"ContainerStarted","Data":"1e998868fc2ee6f2a0a2f21de3e0d1eda34886eb63091e0aefdaa3309cc21d8a"} Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.138316 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.138337 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.150449 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.150524 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.150586 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.151431 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"078c8db152d306f847e249781287a9e53d5a972cc64796b1689df8c3d5c083f5"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.151506 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://078c8db152d306f847e249781287a9e53d5a972cc64796b1689df8c3d5c083f5" gracePeriod=600 Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.174521 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:29 crc kubenswrapper[4960]: I1002 07:26:29.203659 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" podStartSLOduration=8.203636527 podStartE2EDuration="8.203636527s" podCreationTimestamp="2025-10-02 07:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:26:29.165847499 +0000 UTC m=+610.197793786" watchObservedRunningTime="2025-10-02 07:26:29.203636527 +0000 UTC m=+610.235582814" Oct 02 07:26:30 crc kubenswrapper[4960]: I1002 07:26:30.147204 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="078c8db152d306f847e249781287a9e53d5a972cc64796b1689df8c3d5c083f5" exitCode=0 Oct 02 07:26:30 crc kubenswrapper[4960]: I1002 07:26:30.147305 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"078c8db152d306f847e249781287a9e53d5a972cc64796b1689df8c3d5c083f5"} Oct 02 07:26:30 crc kubenswrapper[4960]: I1002 07:26:30.147644 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"856ec690ae49b3ae3dbe5a23fd799e7b1fd275afa9ab067d75708fa279072617"} Oct 02 07:26:30 crc kubenswrapper[4960]: I1002 07:26:30.147679 4960 scope.go:117] "RemoveContainer" containerID="e05a2b35123183dd07d94c3c5aa895e60ba41f5e31b710be48e94c02186f6ea1" Oct 02 07:26:30 crc kubenswrapper[4960]: I1002 07:26:30.148303 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:30 crc kubenswrapper[4960]: I1002 07:26:30.185048 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:26:36 crc kubenswrapper[4960]: I1002 07:26:36.331334 4960 scope.go:117] "RemoveContainer" containerID="58bee76173a69e8b6bf08b9a1abd3626ec971618f989710817a49de54f5243e2" Oct 02 07:26:36 crc kubenswrapper[4960]: E1002 07:26:36.332388 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2dqxx_openshift-multus(c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e)\"" pod="openshift-multus/multus-2dqxx" podUID="c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e" Oct 02 07:26:48 crc kubenswrapper[4960]: I1002 07:26:48.330327 4960 scope.go:117] "RemoveContainer" containerID="58bee76173a69e8b6bf08b9a1abd3626ec971618f989710817a49de54f5243e2" Oct 02 07:26:49 crc kubenswrapper[4960]: I1002 07:26:49.295913 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/2.log" Oct 02 07:26:49 crc kubenswrapper[4960]: I1002 07:26:49.297057 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/1.log" Oct 02 07:26:49 crc kubenswrapper[4960]: I1002 07:26:49.297123 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2dqxx" event={"ID":"c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e","Type":"ContainerStarted","Data":"c6f1e2ee6a057e3b3adf8b5682ae94b369069b237f6bd3c14cec68748d527599"} Oct 02 07:26:52 crc kubenswrapper[4960]: I1002 07:26:52.109500 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h795z" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.497137 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b"] Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.499302 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.506016 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.517580 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b"] Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.613249 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.613368 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxnwf\" (UniqueName: \"kubernetes.io/projected/cc5b1383-a1a0-4e12-908e-241ef19995ff-kube-api-access-fxnwf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.613581 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.714726 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.714885 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.714930 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxnwf\" (UniqueName: \"kubernetes.io/projected/cc5b1383-a1a0-4e12-908e-241ef19995ff-kube-api-access-fxnwf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.715652 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.715833 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.743451 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxnwf\" (UniqueName: \"kubernetes.io/projected/cc5b1383-a1a0-4e12-908e-241ef19995ff-kube-api-access-fxnwf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:04 crc kubenswrapper[4960]: I1002 07:27:04.819437 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:05 crc kubenswrapper[4960]: I1002 07:27:05.171606 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b"] Oct 02 07:27:05 crc kubenswrapper[4960]: I1002 07:27:05.415389 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" event={"ID":"cc5b1383-a1a0-4e12-908e-241ef19995ff","Type":"ContainerStarted","Data":"e2f3dcbd26891f8d15f875634c7dd19f01eed29a8d4b09f8d1030b83b6f13ada"} Oct 02 07:27:05 crc kubenswrapper[4960]: I1002 07:27:05.415851 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" event={"ID":"cc5b1383-a1a0-4e12-908e-241ef19995ff","Type":"ContainerStarted","Data":"738e9942216b068d1d72745f760ee383b0820124d4868612ca2d164d47704c33"} Oct 02 07:27:06 crc kubenswrapper[4960]: I1002 07:27:06.426211 4960 generic.go:334] "Generic (PLEG): container finished" podID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerID="e2f3dcbd26891f8d15f875634c7dd19f01eed29a8d4b09f8d1030b83b6f13ada" exitCode=0 Oct 02 07:27:06 crc kubenswrapper[4960]: I1002 07:27:06.426291 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" event={"ID":"cc5b1383-a1a0-4e12-908e-241ef19995ff","Type":"ContainerDied","Data":"e2f3dcbd26891f8d15f875634c7dd19f01eed29a8d4b09f8d1030b83b6f13ada"} Oct 02 07:27:08 crc kubenswrapper[4960]: I1002 07:27:08.442422 4960 generic.go:334] "Generic (PLEG): container finished" podID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerID="a950cf120ef250c46e8616aa6531b3c8d79d08b4443a300bd443b51a5acf744e" exitCode=0 Oct 02 07:27:08 crc kubenswrapper[4960]: I1002 07:27:08.442506 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" event={"ID":"cc5b1383-a1a0-4e12-908e-241ef19995ff","Type":"ContainerDied","Data":"a950cf120ef250c46e8616aa6531b3c8d79d08b4443a300bd443b51a5acf744e"} Oct 02 07:27:08 crc kubenswrapper[4960]: E1002 07:27:08.806663 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc5b1383_a1a0_4e12_908e_241ef19995ff.slice/crio-conmon-c1b3d78b6bbab7019da1b2a909cbfc4abf829eebc4d1344418f1756585fd0dde.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc5b1383_a1a0_4e12_908e_241ef19995ff.slice/crio-c1b3d78b6bbab7019da1b2a909cbfc4abf829eebc4d1344418f1756585fd0dde.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:27:09 crc kubenswrapper[4960]: I1002 07:27:09.453226 4960 generic.go:334] "Generic (PLEG): container finished" podID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerID="c1b3d78b6bbab7019da1b2a909cbfc4abf829eebc4d1344418f1756585fd0dde" exitCode=0 Oct 02 07:27:09 crc kubenswrapper[4960]: I1002 07:27:09.453330 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" event={"ID":"cc5b1383-a1a0-4e12-908e-241ef19995ff","Type":"ContainerDied","Data":"c1b3d78b6bbab7019da1b2a909cbfc4abf829eebc4d1344418f1756585fd0dde"} Oct 02 07:27:10 crc kubenswrapper[4960]: I1002 07:27:10.798796 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:10 crc kubenswrapper[4960]: I1002 07:27:10.909389 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-util\") pod \"cc5b1383-a1a0-4e12-908e-241ef19995ff\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " Oct 02 07:27:10 crc kubenswrapper[4960]: I1002 07:27:10.909481 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-bundle\") pod \"cc5b1383-a1a0-4e12-908e-241ef19995ff\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " Oct 02 07:27:10 crc kubenswrapper[4960]: I1002 07:27:10.909703 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxnwf\" (UniqueName: \"kubernetes.io/projected/cc5b1383-a1a0-4e12-908e-241ef19995ff-kube-api-access-fxnwf\") pod \"cc5b1383-a1a0-4e12-908e-241ef19995ff\" (UID: \"cc5b1383-a1a0-4e12-908e-241ef19995ff\") " Oct 02 07:27:10 crc kubenswrapper[4960]: I1002 07:27:10.910457 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-bundle" (OuterVolumeSpecName: "bundle") pod "cc5b1383-a1a0-4e12-908e-241ef19995ff" (UID: "cc5b1383-a1a0-4e12-908e-241ef19995ff"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:27:10 crc kubenswrapper[4960]: I1002 07:27:10.920081 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc5b1383-a1a0-4e12-908e-241ef19995ff-kube-api-access-fxnwf" (OuterVolumeSpecName: "kube-api-access-fxnwf") pod "cc5b1383-a1a0-4e12-908e-241ef19995ff" (UID: "cc5b1383-a1a0-4e12-908e-241ef19995ff"). InnerVolumeSpecName "kube-api-access-fxnwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:27:10 crc kubenswrapper[4960]: I1002 07:27:10.932538 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-util" (OuterVolumeSpecName: "util") pod "cc5b1383-a1a0-4e12-908e-241ef19995ff" (UID: "cc5b1383-a1a0-4e12-908e-241ef19995ff"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:27:11 crc kubenswrapper[4960]: I1002 07:27:11.011196 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxnwf\" (UniqueName: \"kubernetes.io/projected/cc5b1383-a1a0-4e12-908e-241ef19995ff-kube-api-access-fxnwf\") on node \"crc\" DevicePath \"\"" Oct 02 07:27:11 crc kubenswrapper[4960]: I1002 07:27:11.011233 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-util\") on node \"crc\" DevicePath \"\"" Oct 02 07:27:11 crc kubenswrapper[4960]: I1002 07:27:11.011245 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5b1383-a1a0-4e12-908e-241ef19995ff-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:27:11 crc kubenswrapper[4960]: I1002 07:27:11.487318 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" event={"ID":"cc5b1383-a1a0-4e12-908e-241ef19995ff","Type":"ContainerDied","Data":"738e9942216b068d1d72745f760ee383b0820124d4868612ca2d164d47704c33"} Oct 02 07:27:11 crc kubenswrapper[4960]: I1002 07:27:11.487776 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="738e9942216b068d1d72745f760ee383b0820124d4868612ca2d164d47704c33" Oct 02 07:27:11 crc kubenswrapper[4960]: I1002 07:27:11.487454 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.088339 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s"] Oct 02 07:27:16 crc kubenswrapper[4960]: E1002 07:27:16.088898 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerName="extract" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.088918 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerName="extract" Oct 02 07:27:16 crc kubenswrapper[4960]: E1002 07:27:16.088937 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerName="pull" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.088943 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerName="pull" Oct 02 07:27:16 crc kubenswrapper[4960]: E1002 07:27:16.088961 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerName="util" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.088967 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerName="util" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.089070 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc5b1383-a1a0-4e12-908e-241ef19995ff" containerName="extract" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.089486 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.091861 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jqcls" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.092175 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.093566 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.106515 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s"] Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.183696 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2twjm\" (UniqueName: \"kubernetes.io/projected/e1c55993-cf88-4aac-a665-92e26a69a841-kube-api-access-2twjm\") pod \"nmstate-operator-858ddd8f98-vpq7s\" (UID: \"e1c55993-cf88-4aac-a665-92e26a69a841\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.284663 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2twjm\" (UniqueName: \"kubernetes.io/projected/e1c55993-cf88-4aac-a665-92e26a69a841-kube-api-access-2twjm\") pod \"nmstate-operator-858ddd8f98-vpq7s\" (UID: \"e1c55993-cf88-4aac-a665-92e26a69a841\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.311193 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2twjm\" (UniqueName: \"kubernetes.io/projected/e1c55993-cf88-4aac-a665-92e26a69a841-kube-api-access-2twjm\") pod \"nmstate-operator-858ddd8f98-vpq7s\" (UID: \"e1c55993-cf88-4aac-a665-92e26a69a841\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.410476 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" Oct 02 07:27:16 crc kubenswrapper[4960]: I1002 07:27:16.645428 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s"] Oct 02 07:27:16 crc kubenswrapper[4960]: W1002 07:27:16.657563 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1c55993_cf88_4aac_a665_92e26a69a841.slice/crio-5c2dfd5c5e120897c4710ac45a60bf648b666b417821d3a1e4d784776c7671c9 WatchSource:0}: Error finding container 5c2dfd5c5e120897c4710ac45a60bf648b666b417821d3a1e4d784776c7671c9: Status 404 returned error can't find the container with id 5c2dfd5c5e120897c4710ac45a60bf648b666b417821d3a1e4d784776c7671c9 Oct 02 07:27:17 crc kubenswrapper[4960]: I1002 07:27:17.531007 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" event={"ID":"e1c55993-cf88-4aac-a665-92e26a69a841","Type":"ContainerStarted","Data":"5c2dfd5c5e120897c4710ac45a60bf648b666b417821d3a1e4d784776c7671c9"} Oct 02 07:27:19 crc kubenswrapper[4960]: I1002 07:27:19.548547 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" event={"ID":"e1c55993-cf88-4aac-a665-92e26a69a841","Type":"ContainerStarted","Data":"993fac24952e7b3b089bc806b8471c3d17096937c7db20c9473cb869bc5f65f9"} Oct 02 07:27:19 crc kubenswrapper[4960]: I1002 07:27:19.570944 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vpq7s" podStartSLOduration=1.342657209 podStartE2EDuration="3.570917055s" podCreationTimestamp="2025-10-02 07:27:16 +0000 UTC" firstStartedPulling="2025-10-02 07:27:16.660835343 +0000 UTC m=+657.692781640" lastFinishedPulling="2025-10-02 07:27:18.889095189 +0000 UTC m=+659.921041486" observedRunningTime="2025-10-02 07:27:19.570330525 +0000 UTC m=+660.602276842" watchObservedRunningTime="2025-10-02 07:27:19.570917055 +0000 UTC m=+660.602863372" Oct 02 07:27:20 crc kubenswrapper[4960]: I1002 07:27:20.498297 4960 scope.go:117] "RemoveContainer" containerID="f54e5ed1e8592325e2961e104cb502c319443eaa7b8db1d48b11fbd87c32cdd3" Oct 02 07:27:20 crc kubenswrapper[4960]: I1002 07:27:20.558104 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2dqxx_c96f3ba3-6aa3-4f51-a8cc-1b21ef16f29e/kube-multus/2.log" Oct 02 07:27:24 crc kubenswrapper[4960]: I1002 07:27:24.982673 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr"] Oct 02 07:27:24 crc kubenswrapper[4960]: I1002 07:27:24.984312 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" Oct 02 07:27:24 crc kubenswrapper[4960]: I1002 07:27:24.988436 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-zf452" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.000793 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.002003 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.003867 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.004289 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.042621 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-r24b6"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.043907 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.052181 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.115315 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-dbus-socket\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.115369 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx9z9\" (UniqueName: \"kubernetes.io/projected/3242a27c-0911-439a-b346-0f921d212cb8-kube-api-access-tx9z9\") pod \"nmstate-metrics-fdff9cb8d-77kzr\" (UID: \"3242a27c-0911-439a-b346-0f921d212cb8\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.115416 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr4wx\" (UniqueName: \"kubernetes.io/projected/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-kube-api-access-wr4wx\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.115434 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-cjznw\" (UID: \"b1cd06a6-a984-40d3-bd87-e81eedce2c7c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.115470 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-ovs-socket\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.115614 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-nmstate-lock\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.115633 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtqzv\" (UniqueName: \"kubernetes.io/projected/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-kube-api-access-vtqzv\") pod \"nmstate-webhook-6cdbc54649-cjznw\" (UID: \"b1cd06a6-a984-40d3-bd87-e81eedce2c7c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.154242 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.155022 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.157554 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.157666 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.158298 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dt6cl" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.171146 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.216694 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-dbus-socket\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.216808 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx9z9\" (UniqueName: \"kubernetes.io/projected/3242a27c-0911-439a-b346-0f921d212cb8-kube-api-access-tx9z9\") pod \"nmstate-metrics-fdff9cb8d-77kzr\" (UID: \"3242a27c-0911-439a-b346-0f921d212cb8\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.217143 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-dbus-socket\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.217450 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr4wx\" (UniqueName: \"kubernetes.io/projected/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-kube-api-access-wr4wx\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.217481 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-cjznw\" (UID: \"b1cd06a6-a984-40d3-bd87-e81eedce2c7c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: E1002 07:27:25.217785 4960 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 02 07:27:25 crc kubenswrapper[4960]: E1002 07:27:25.217850 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-tls-key-pair podName:b1cd06a6-a984-40d3-bd87-e81eedce2c7c nodeName:}" failed. No retries permitted until 2025-10-02 07:27:25.717828257 +0000 UTC m=+666.749774544 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-tls-key-pair") pod "nmstate-webhook-6cdbc54649-cjznw" (UID: "b1cd06a6-a984-40d3-bd87-e81eedce2c7c") : secret "openshift-nmstate-webhook" not found Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218112 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcf6c\" (UniqueName: \"kubernetes.io/projected/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-kube-api-access-tcf6c\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218178 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-ovs-socket\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218243 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-ovs-socket\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218404 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218437 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218494 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtqzv\" (UniqueName: \"kubernetes.io/projected/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-kube-api-access-vtqzv\") pod \"nmstate-webhook-6cdbc54649-cjznw\" (UID: \"b1cd06a6-a984-40d3-bd87-e81eedce2c7c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218557 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-nmstate-lock\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.218812 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-nmstate-lock\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.238230 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr4wx\" (UniqueName: \"kubernetes.io/projected/bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67-kube-api-access-wr4wx\") pod \"nmstate-handler-r24b6\" (UID: \"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67\") " pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.238315 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx9z9\" (UniqueName: \"kubernetes.io/projected/3242a27c-0911-439a-b346-0f921d212cb8-kube-api-access-tx9z9\") pod \"nmstate-metrics-fdff9cb8d-77kzr\" (UID: \"3242a27c-0911-439a-b346-0f921d212cb8\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.251564 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtqzv\" (UniqueName: \"kubernetes.io/projected/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-kube-api-access-vtqzv\") pod \"nmstate-webhook-6cdbc54649-cjznw\" (UID: \"b1cd06a6-a984-40d3-bd87-e81eedce2c7c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.320149 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcf6c\" (UniqueName: \"kubernetes.io/projected/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-kube-api-access-tcf6c\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.320211 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.320229 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.321191 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.325596 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.334627 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6d6c4458c4-hv54l"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.335660 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.336304 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.348846 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d6c4458c4-hv54l"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.357644 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcf6c\" (UniqueName: \"kubernetes.io/projected/d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb-kube-api-access-tcf6c\") pod \"nmstate-console-plugin-6b874cbd85-k8xfq\" (UID: \"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.367914 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:25 crc kubenswrapper[4960]: W1002 07:27:25.395014 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf2ea1e7_4cf6_4bc3_be78_37490d9ffc67.slice/crio-ec113a9cc6b8391e49f53b1280976edae9f74177b8608236dd05c1974d9fa6c1 WatchSource:0}: Error finding container ec113a9cc6b8391e49f53b1280976edae9f74177b8608236dd05c1974d9fa6c1: Status 404 returned error can't find the container with id ec113a9cc6b8391e49f53b1280976edae9f74177b8608236dd05c1974d9fa6c1 Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.422058 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-serving-cert\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.422125 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb7mp\" (UniqueName: \"kubernetes.io/projected/5f905791-e2d4-46bb-9e7e-fbffa36ba114-kube-api-access-pb7mp\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.422223 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-service-ca\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.422258 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-trusted-ca-bundle\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.422284 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-oauth-serving-cert\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.422307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-oauth-config\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.422541 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-config\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.470551 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.524073 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-config\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.524137 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-serving-cert\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.524191 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb7mp\" (UniqueName: \"kubernetes.io/projected/5f905791-e2d4-46bb-9e7e-fbffa36ba114-kube-api-access-pb7mp\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.524221 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-service-ca\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.524262 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-trusted-ca-bundle\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.524282 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-oauth-serving-cert\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.524302 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-oauth-config\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.525281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-config\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.526222 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-trusted-ca-bundle\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.526354 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-service-ca\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.526723 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5f905791-e2d4-46bb-9e7e-fbffa36ba114-oauth-serving-cert\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.528460 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-serving-cert\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.528813 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5f905791-e2d4-46bb-9e7e-fbffa36ba114-console-oauth-config\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.547997 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb7mp\" (UniqueName: \"kubernetes.io/projected/5f905791-e2d4-46bb-9e7e-fbffa36ba114-kube-api-access-pb7mp\") pod \"console-6d6c4458c4-hv54l\" (UID: \"5f905791-e2d4-46bb-9e7e-fbffa36ba114\") " pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.553892 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr"] Oct 02 07:27:25 crc kubenswrapper[4960]: W1002 07:27:25.565427 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3242a27c_0911_439a_b346_0f921d212cb8.slice/crio-cb0d20dadc301862ed465ffe43ff4b618ddc17790ec6517fbf42ed52bf4b31f2 WatchSource:0}: Error finding container cb0d20dadc301862ed465ffe43ff4b618ddc17790ec6517fbf42ed52bf4b31f2: Status 404 returned error can't find the container with id cb0d20dadc301862ed465ffe43ff4b618ddc17790ec6517fbf42ed52bf4b31f2 Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.596428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" event={"ID":"3242a27c-0911-439a-b346-0f921d212cb8","Type":"ContainerStarted","Data":"cb0d20dadc301862ed465ffe43ff4b618ddc17790ec6517fbf42ed52bf4b31f2"} Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.597541 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-r24b6" event={"ID":"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67","Type":"ContainerStarted","Data":"ec113a9cc6b8391e49f53b1280976edae9f74177b8608236dd05c1974d9fa6c1"} Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.714516 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.722161 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq"] Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.727251 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-cjznw\" (UID: \"b1cd06a6-a984-40d3-bd87-e81eedce2c7c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.732354 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b1cd06a6-a984-40d3-bd87-e81eedce2c7c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-cjznw\" (UID: \"b1cd06a6-a984-40d3-bd87-e81eedce2c7c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:25 crc kubenswrapper[4960]: W1002 07:27:25.735189 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3bfd63a_eb72_47a8_a5d1_ca68b6fd40fb.slice/crio-d49ce4ae421898b165a611f5c471b77edd742a37c422709035bd5330ba6a66d8 WatchSource:0}: Error finding container d49ce4ae421898b165a611f5c471b77edd742a37c422709035bd5330ba6a66d8: Status 404 returned error can't find the container with id d49ce4ae421898b165a611f5c471b77edd742a37c422709035bd5330ba6a66d8 Oct 02 07:27:25 crc kubenswrapper[4960]: I1002 07:27:25.945113 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:26 crc kubenswrapper[4960]: I1002 07:27:26.124405 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d6c4458c4-hv54l"] Oct 02 07:27:26 crc kubenswrapper[4960]: W1002 07:27:26.133382 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f905791_e2d4_46bb_9e7e_fbffa36ba114.slice/crio-6f1f6e78695c5f317cedc4746f1e9196db3e646fff3d117623a77f437bb9da21 WatchSource:0}: Error finding container 6f1f6e78695c5f317cedc4746f1e9196db3e646fff3d117623a77f437bb9da21: Status 404 returned error can't find the container with id 6f1f6e78695c5f317cedc4746f1e9196db3e646fff3d117623a77f437bb9da21 Oct 02 07:27:26 crc kubenswrapper[4960]: I1002 07:27:26.169279 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw"] Oct 02 07:27:26 crc kubenswrapper[4960]: W1002 07:27:26.180588 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1cd06a6_a984_40d3_bd87_e81eedce2c7c.slice/crio-a43fab537f6b5e14d5c1d786c110b77e50f2b1e0e0aba0e63fed6ffa3103335c WatchSource:0}: Error finding container a43fab537f6b5e14d5c1d786c110b77e50f2b1e0e0aba0e63fed6ffa3103335c: Status 404 returned error can't find the container with id a43fab537f6b5e14d5c1d786c110b77e50f2b1e0e0aba0e63fed6ffa3103335c Oct 02 07:27:26 crc kubenswrapper[4960]: I1002 07:27:26.607797 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" event={"ID":"b1cd06a6-a984-40d3-bd87-e81eedce2c7c","Type":"ContainerStarted","Data":"a43fab537f6b5e14d5c1d786c110b77e50f2b1e0e0aba0e63fed6ffa3103335c"} Oct 02 07:27:26 crc kubenswrapper[4960]: I1002 07:27:26.611894 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" event={"ID":"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb","Type":"ContainerStarted","Data":"d49ce4ae421898b165a611f5c471b77edd742a37c422709035bd5330ba6a66d8"} Oct 02 07:27:26 crc kubenswrapper[4960]: I1002 07:27:26.614518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d6c4458c4-hv54l" event={"ID":"5f905791-e2d4-46bb-9e7e-fbffa36ba114","Type":"ContainerStarted","Data":"a719034db44d88d3b439b18ae8d285f8de0b4a0a97d874d4a5a80fcf79da30d9"} Oct 02 07:27:26 crc kubenswrapper[4960]: I1002 07:27:26.614556 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d6c4458c4-hv54l" event={"ID":"5f905791-e2d4-46bb-9e7e-fbffa36ba114","Type":"ContainerStarted","Data":"6f1f6e78695c5f317cedc4746f1e9196db3e646fff3d117623a77f437bb9da21"} Oct 02 07:27:26 crc kubenswrapper[4960]: I1002 07:27:26.641901 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6d6c4458c4-hv54l" podStartSLOduration=1.641870569 podStartE2EDuration="1.641870569s" podCreationTimestamp="2025-10-02 07:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:27:26.635859064 +0000 UTC m=+667.667805371" watchObservedRunningTime="2025-10-02 07:27:26.641870569 +0000 UTC m=+667.673816866" Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.636200 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" event={"ID":"3242a27c-0911-439a-b346-0f921d212cb8","Type":"ContainerStarted","Data":"13c54d791fb84a8d18c861449923dc0216effed1c15d80366154e608d1f51c7a"} Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.639202 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-r24b6" event={"ID":"bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67","Type":"ContainerStarted","Data":"e382290930ce3e70c95fa5d1f5d4a5610e7dc23b981817a6ee4053cea66a593c"} Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.639343 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.641798 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" event={"ID":"b1cd06a6-a984-40d3-bd87-e81eedce2c7c","Type":"ContainerStarted","Data":"45c4ce03e629997fbb88b2105c26346c7b0e7c099184869b0593ab7f3bcf71f4"} Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.641911 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.643507 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" event={"ID":"d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb","Type":"ContainerStarted","Data":"543447fa4900e6c46549c542428a601b02bd4dd16194140041a2c2ec6b4665ce"} Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.656260 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-r24b6" podStartSLOduration=2.566743058 podStartE2EDuration="5.656233474s" podCreationTimestamp="2025-10-02 07:27:24 +0000 UTC" firstStartedPulling="2025-10-02 07:27:25.40260795 +0000 UTC m=+666.434554237" lastFinishedPulling="2025-10-02 07:27:28.492098366 +0000 UTC m=+669.524044653" observedRunningTime="2025-10-02 07:27:29.655282357 +0000 UTC m=+670.687228704" watchObservedRunningTime="2025-10-02 07:27:29.656233474 +0000 UTC m=+670.688179801" Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.681235 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" podStartSLOduration=3.371710857 podStartE2EDuration="5.681212659s" podCreationTimestamp="2025-10-02 07:27:24 +0000 UTC" firstStartedPulling="2025-10-02 07:27:26.189163179 +0000 UTC m=+667.221109456" lastFinishedPulling="2025-10-02 07:27:28.498664961 +0000 UTC m=+669.530611258" observedRunningTime="2025-10-02 07:27:29.678456287 +0000 UTC m=+670.710402664" watchObservedRunningTime="2025-10-02 07:27:29.681212659 +0000 UTC m=+670.713158946" Oct 02 07:27:29 crc kubenswrapper[4960]: I1002 07:27:29.699566 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-k8xfq" podStartSLOduration=1.943989514 podStartE2EDuration="4.699540488s" podCreationTimestamp="2025-10-02 07:27:25 +0000 UTC" firstStartedPulling="2025-10-02 07:27:25.737589392 +0000 UTC m=+666.769535679" lastFinishedPulling="2025-10-02 07:27:28.493140366 +0000 UTC m=+669.525086653" observedRunningTime="2025-10-02 07:27:29.697603541 +0000 UTC m=+670.729549848" watchObservedRunningTime="2025-10-02 07:27:29.699540488 +0000 UTC m=+670.731486795" Oct 02 07:27:32 crc kubenswrapper[4960]: I1002 07:27:32.666237 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" event={"ID":"3242a27c-0911-439a-b346-0f921d212cb8","Type":"ContainerStarted","Data":"7c5b9d6e61a33ef5044cfa3633395e0fcd474d0e22a514e91357249a6966ea6a"} Oct 02 07:27:32 crc kubenswrapper[4960]: I1002 07:27:32.695052 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-77kzr" podStartSLOduration=1.979285478 podStartE2EDuration="8.695020785s" podCreationTimestamp="2025-10-02 07:27:24 +0000 UTC" firstStartedPulling="2025-10-02 07:27:25.567898995 +0000 UTC m=+666.599845282" lastFinishedPulling="2025-10-02 07:27:32.283634302 +0000 UTC m=+673.315580589" observedRunningTime="2025-10-02 07:27:32.687725717 +0000 UTC m=+673.719672034" watchObservedRunningTime="2025-10-02 07:27:32.695020785 +0000 UTC m=+673.726967112" Oct 02 07:27:35 crc kubenswrapper[4960]: I1002 07:27:35.413479 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-r24b6" Oct 02 07:27:35 crc kubenswrapper[4960]: I1002 07:27:35.715403 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:35 crc kubenswrapper[4960]: I1002 07:27:35.715489 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:35 crc kubenswrapper[4960]: I1002 07:27:35.721273 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:36 crc kubenswrapper[4960]: I1002 07:27:36.705165 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6d6c4458c4-hv54l" Oct 02 07:27:36 crc kubenswrapper[4960]: I1002 07:27:36.791555 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vz5gr"] Oct 02 07:27:45 crc kubenswrapper[4960]: I1002 07:27:45.957370 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-cjznw" Oct 02 07:28:01 crc kubenswrapper[4960]: I1002 07:28:01.853803 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vz5gr" podUID="a5911cf0-70ba-44a3-8c49-4cd66cca73b7" containerName="console" containerID="cri-o://5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22" gracePeriod=15 Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.258599 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vz5gr_a5911cf0-70ba-44a3-8c49-4cd66cca73b7/console/0.log" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.259008 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.329498 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-oauth-config\") pod \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.329568 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-serving-cert\") pod \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.329620 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j22gj\" (UniqueName: \"kubernetes.io/projected/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-kube-api-access-j22gj\") pod \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.329673 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-trusted-ca-bundle\") pod \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.329709 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-config\") pod \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.329768 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-oauth-serving-cert\") pod \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.329810 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-service-ca\") pod \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\" (UID: \"a5911cf0-70ba-44a3-8c49-4cd66cca73b7\") " Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.330640 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a5911cf0-70ba-44a3-8c49-4cd66cca73b7" (UID: "a5911cf0-70ba-44a3-8c49-4cd66cca73b7"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.331064 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-service-ca" (OuterVolumeSpecName: "service-ca") pod "a5911cf0-70ba-44a3-8c49-4cd66cca73b7" (UID: "a5911cf0-70ba-44a3-8c49-4cd66cca73b7"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.331161 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-config" (OuterVolumeSpecName: "console-config") pod "a5911cf0-70ba-44a3-8c49-4cd66cca73b7" (UID: "a5911cf0-70ba-44a3-8c49-4cd66cca73b7"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.331202 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a5911cf0-70ba-44a3-8c49-4cd66cca73b7" (UID: "a5911cf0-70ba-44a3-8c49-4cd66cca73b7"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.338897 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-kube-api-access-j22gj" (OuterVolumeSpecName: "kube-api-access-j22gj") pod "a5911cf0-70ba-44a3-8c49-4cd66cca73b7" (UID: "a5911cf0-70ba-44a3-8c49-4cd66cca73b7"). InnerVolumeSpecName "kube-api-access-j22gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.339015 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a5911cf0-70ba-44a3-8c49-4cd66cca73b7" (UID: "a5911cf0-70ba-44a3-8c49-4cd66cca73b7"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.341321 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a5911cf0-70ba-44a3-8c49-4cd66cca73b7" (UID: "a5911cf0-70ba-44a3-8c49-4cd66cca73b7"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.431911 4960 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.431948 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.431957 4960 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.431967 4960 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.431990 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j22gj\" (UniqueName: \"kubernetes.io/projected/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-kube-api-access-j22gj\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.432000 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.432009 4960 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a5911cf0-70ba-44a3-8c49-4cd66cca73b7-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.913508 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vz5gr_a5911cf0-70ba-44a3-8c49-4cd66cca73b7/console/0.log" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.913557 4960 generic.go:334] "Generic (PLEG): container finished" podID="a5911cf0-70ba-44a3-8c49-4cd66cca73b7" containerID="5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22" exitCode=2 Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.913595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vz5gr" event={"ID":"a5911cf0-70ba-44a3-8c49-4cd66cca73b7","Type":"ContainerDied","Data":"5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22"} Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.913625 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vz5gr" event={"ID":"a5911cf0-70ba-44a3-8c49-4cd66cca73b7","Type":"ContainerDied","Data":"6ef0935cde1ec223e9c4d5726645648e4cf73f5c2fbd2fd4fdffcd8405bd564f"} Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.913645 4960 scope.go:117] "RemoveContainer" containerID="5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.913767 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vz5gr" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.951384 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vz5gr"] Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.951491 4960 scope.go:117] "RemoveContainer" containerID="5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22" Oct 02 07:28:02 crc kubenswrapper[4960]: E1002 07:28:02.953166 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22\": container with ID starting with 5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22 not found: ID does not exist" containerID="5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.953244 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22"} err="failed to get container status \"5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22\": rpc error: code = NotFound desc = could not find container \"5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22\": container with ID starting with 5a017a47eac427d3b3328677567d211511f092b16ddbfb7414f2c0bc25e68a22 not found: ID does not exist" Oct 02 07:28:02 crc kubenswrapper[4960]: I1002 07:28:02.963767 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vz5gr"] Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.519731 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n"] Oct 02 07:28:03 crc kubenswrapper[4960]: E1002 07:28:03.520030 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5911cf0-70ba-44a3-8c49-4cd66cca73b7" containerName="console" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.520053 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5911cf0-70ba-44a3-8c49-4cd66cca73b7" containerName="console" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.520193 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5911cf0-70ba-44a3-8c49-4cd66cca73b7" containerName="console" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.521134 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.523254 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.530646 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n"] Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.649747 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lgd6\" (UniqueName: \"kubernetes.io/projected/d744d009-5615-41b8-a17c-0e290f4d73c6-kube-api-access-5lgd6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.649826 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.650178 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.751587 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.751700 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lgd6\" (UniqueName: \"kubernetes.io/projected/d744d009-5615-41b8-a17c-0e290f4d73c6-kube-api-access-5lgd6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.751747 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.752361 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.752404 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.769377 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lgd6\" (UniqueName: \"kubernetes.io/projected/d744d009-5615-41b8-a17c-0e290f4d73c6-kube-api-access-5lgd6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:03 crc kubenswrapper[4960]: I1002 07:28:03.843374 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:04 crc kubenswrapper[4960]: I1002 07:28:04.039597 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n"] Oct 02 07:28:04 crc kubenswrapper[4960]: I1002 07:28:04.340662 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5911cf0-70ba-44a3-8c49-4cd66cca73b7" path="/var/lib/kubelet/pods/a5911cf0-70ba-44a3-8c49-4cd66cca73b7/volumes" Oct 02 07:28:04 crc kubenswrapper[4960]: I1002 07:28:04.936150 4960 generic.go:334] "Generic (PLEG): container finished" podID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerID="0ba773280f0566ac332375ae52e8f63b8d31744ad0a2cfb2a14f6321809dabd7" exitCode=0 Oct 02 07:28:04 crc kubenswrapper[4960]: I1002 07:28:04.936201 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" event={"ID":"d744d009-5615-41b8-a17c-0e290f4d73c6","Type":"ContainerDied","Data":"0ba773280f0566ac332375ae52e8f63b8d31744ad0a2cfb2a14f6321809dabd7"} Oct 02 07:28:04 crc kubenswrapper[4960]: I1002 07:28:04.936230 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" event={"ID":"d744d009-5615-41b8-a17c-0e290f4d73c6","Type":"ContainerStarted","Data":"df34ade4ed7344559375688fad7cc6b3af988bc8754f7a661ee0f43854a7ca96"} Oct 02 07:28:06 crc kubenswrapper[4960]: I1002 07:28:06.962208 4960 generic.go:334] "Generic (PLEG): container finished" podID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerID="5798c0bbe00aa96b554b5ee07e3585b2bdfd0c0032a4e995b6a2d1d5f082ef01" exitCode=0 Oct 02 07:28:06 crc kubenswrapper[4960]: I1002 07:28:06.962415 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" event={"ID":"d744d009-5615-41b8-a17c-0e290f4d73c6","Type":"ContainerDied","Data":"5798c0bbe00aa96b554b5ee07e3585b2bdfd0c0032a4e995b6a2d1d5f082ef01"} Oct 02 07:28:07 crc kubenswrapper[4960]: I1002 07:28:07.984650 4960 generic.go:334] "Generic (PLEG): container finished" podID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerID="41fe9c81c064828cecd5b15a1def643a2e70136378c23f3d6f561c320b59e7e3" exitCode=0 Oct 02 07:28:07 crc kubenswrapper[4960]: I1002 07:28:07.984716 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" event={"ID":"d744d009-5615-41b8-a17c-0e290f4d73c6","Type":"ContainerDied","Data":"41fe9c81c064828cecd5b15a1def643a2e70136378c23f3d6f561c320b59e7e3"} Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.310737 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.445426 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lgd6\" (UniqueName: \"kubernetes.io/projected/d744d009-5615-41b8-a17c-0e290f4d73c6-kube-api-access-5lgd6\") pod \"d744d009-5615-41b8-a17c-0e290f4d73c6\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.445552 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-util\") pod \"d744d009-5615-41b8-a17c-0e290f4d73c6\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.445589 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-bundle\") pod \"d744d009-5615-41b8-a17c-0e290f4d73c6\" (UID: \"d744d009-5615-41b8-a17c-0e290f4d73c6\") " Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.448570 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-bundle" (OuterVolumeSpecName: "bundle") pod "d744d009-5615-41b8-a17c-0e290f4d73c6" (UID: "d744d009-5615-41b8-a17c-0e290f4d73c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.455667 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d744d009-5615-41b8-a17c-0e290f4d73c6-kube-api-access-5lgd6" (OuterVolumeSpecName: "kube-api-access-5lgd6") pod "d744d009-5615-41b8-a17c-0e290f4d73c6" (UID: "d744d009-5615-41b8-a17c-0e290f4d73c6"). InnerVolumeSpecName "kube-api-access-5lgd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.548037 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.548082 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lgd6\" (UniqueName: \"kubernetes.io/projected/d744d009-5615-41b8-a17c-0e290f4d73c6-kube-api-access-5lgd6\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.755836 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-util" (OuterVolumeSpecName: "util") pod "d744d009-5615-41b8-a17c-0e290f4d73c6" (UID: "d744d009-5615-41b8-a17c-0e290f4d73c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:28:09 crc kubenswrapper[4960]: I1002 07:28:09.851592 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d744d009-5615-41b8-a17c-0e290f4d73c6-util\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:10 crc kubenswrapper[4960]: I1002 07:28:10.004125 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" event={"ID":"d744d009-5615-41b8-a17c-0e290f4d73c6","Type":"ContainerDied","Data":"df34ade4ed7344559375688fad7cc6b3af988bc8754f7a661ee0f43854a7ca96"} Oct 02 07:28:10 crc kubenswrapper[4960]: I1002 07:28:10.004169 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df34ade4ed7344559375688fad7cc6b3af988bc8754f7a661ee0f43854a7ca96" Oct 02 07:28:10 crc kubenswrapper[4960]: I1002 07:28:10.004217 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.229068 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg"] Oct 02 07:28:21 crc kubenswrapper[4960]: E1002 07:28:21.229740 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerName="pull" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.229751 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerName="pull" Oct 02 07:28:21 crc kubenswrapper[4960]: E1002 07:28:21.229768 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerName="util" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.229774 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerName="util" Oct 02 07:28:21 crc kubenswrapper[4960]: E1002 07:28:21.229790 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerName="extract" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.229795 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerName="extract" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.229890 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d744d009-5615-41b8-a17c-0e290f4d73c6" containerName="extract" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.230312 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.234996 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.235777 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.235817 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9qmsn" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.235922 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.239100 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.262500 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg"] Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.326381 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b96bb98b-d48b-47d6-8357-0c69cd9052c1-webhook-cert\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.326452 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82n7d\" (UniqueName: \"kubernetes.io/projected/b96bb98b-d48b-47d6-8357-0c69cd9052c1-kube-api-access-82n7d\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.326518 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b96bb98b-d48b-47d6-8357-0c69cd9052c1-apiservice-cert\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.428269 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82n7d\" (UniqueName: \"kubernetes.io/projected/b96bb98b-d48b-47d6-8357-0c69cd9052c1-kube-api-access-82n7d\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.428409 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b96bb98b-d48b-47d6-8357-0c69cd9052c1-apiservice-cert\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.428448 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b96bb98b-d48b-47d6-8357-0c69cd9052c1-webhook-cert\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.439890 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b96bb98b-d48b-47d6-8357-0c69cd9052c1-webhook-cert\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.440509 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b96bb98b-d48b-47d6-8357-0c69cd9052c1-apiservice-cert\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.455750 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82n7d\" (UniqueName: \"kubernetes.io/projected/b96bb98b-d48b-47d6-8357-0c69cd9052c1-kube-api-access-82n7d\") pod \"metallb-operator-controller-manager-cf55b8d8b-rzqpg\" (UID: \"b96bb98b-d48b-47d6-8357-0c69cd9052c1\") " pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.548157 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.619505 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9"] Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.620322 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.623403 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.623836 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-nrnx9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.624231 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.631535 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8d221df4-ed14-4f8d-82db-6326eaa53a3d-apiservice-cert\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.631778 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjl7x\" (UniqueName: \"kubernetes.io/projected/8d221df4-ed14-4f8d-82db-6326eaa53a3d-kube-api-access-gjl7x\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.631927 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8d221df4-ed14-4f8d-82db-6326eaa53a3d-webhook-cert\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.638752 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9"] Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.732795 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjl7x\" (UniqueName: \"kubernetes.io/projected/8d221df4-ed14-4f8d-82db-6326eaa53a3d-kube-api-access-gjl7x\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.733258 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8d221df4-ed14-4f8d-82db-6326eaa53a3d-webhook-cert\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.733295 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8d221df4-ed14-4f8d-82db-6326eaa53a3d-apiservice-cert\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.745892 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8d221df4-ed14-4f8d-82db-6326eaa53a3d-webhook-cert\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.746364 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8d221df4-ed14-4f8d-82db-6326eaa53a3d-apiservice-cert\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.755831 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjl7x\" (UniqueName: \"kubernetes.io/projected/8d221df4-ed14-4f8d-82db-6326eaa53a3d-kube-api-access-gjl7x\") pod \"metallb-operator-webhook-server-db8949b7d-zx4b9\" (UID: \"8d221df4-ed14-4f8d-82db-6326eaa53a3d\") " pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.823893 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg"] Oct 02 07:28:21 crc kubenswrapper[4960]: W1002 07:28:21.839672 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb96bb98b_d48b_47d6_8357_0c69cd9052c1.slice/crio-8c6180b0225e0f08906a0bf25b98b560a0001cdc87ff0c92311d5ebfaf796925 WatchSource:0}: Error finding container 8c6180b0225e0f08906a0bf25b98b560a0001cdc87ff0c92311d5ebfaf796925: Status 404 returned error can't find the container with id 8c6180b0225e0f08906a0bf25b98b560a0001cdc87ff0c92311d5ebfaf796925 Oct 02 07:28:21 crc kubenswrapper[4960]: I1002 07:28:21.942759 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:22 crc kubenswrapper[4960]: I1002 07:28:22.101571 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" event={"ID":"b96bb98b-d48b-47d6-8357-0c69cd9052c1","Type":"ContainerStarted","Data":"8c6180b0225e0f08906a0bf25b98b560a0001cdc87ff0c92311d5ebfaf796925"} Oct 02 07:28:22 crc kubenswrapper[4960]: I1002 07:28:22.237207 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9"] Oct 02 07:28:22 crc kubenswrapper[4960]: W1002 07:28:22.244565 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d221df4_ed14_4f8d_82db_6326eaa53a3d.slice/crio-81a1098c60cdb3de32b5404b1cb5e68e0af39fc2cfa113f5df6a378572d48af8 WatchSource:0}: Error finding container 81a1098c60cdb3de32b5404b1cb5e68e0af39fc2cfa113f5df6a378572d48af8: Status 404 returned error can't find the container with id 81a1098c60cdb3de32b5404b1cb5e68e0af39fc2cfa113f5df6a378572d48af8 Oct 02 07:28:23 crc kubenswrapper[4960]: I1002 07:28:23.117193 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" event={"ID":"8d221df4-ed14-4f8d-82db-6326eaa53a3d","Type":"ContainerStarted","Data":"81a1098c60cdb3de32b5404b1cb5e68e0af39fc2cfa113f5df6a378572d48af8"} Oct 02 07:28:25 crc kubenswrapper[4960]: I1002 07:28:25.145388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" event={"ID":"b96bb98b-d48b-47d6-8357-0c69cd9052c1","Type":"ContainerStarted","Data":"347497a16e70779b937c429adf6ab426100d24080aa19538b87127db88426281"} Oct 02 07:28:25 crc kubenswrapper[4960]: I1002 07:28:25.149507 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:28:28 crc kubenswrapper[4960]: I1002 07:28:28.165754 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" event={"ID":"8d221df4-ed14-4f8d-82db-6326eaa53a3d","Type":"ContainerStarted","Data":"c20ee1799b8b8e85baec5c7546c700a02942fc321952b12d7ceb12f3e35d58f8"} Oct 02 07:28:28 crc kubenswrapper[4960]: I1002 07:28:28.166128 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:28 crc kubenswrapper[4960]: I1002 07:28:28.192904 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" podStartSLOduration=4.159112903 podStartE2EDuration="7.192883547s" podCreationTimestamp="2025-10-02 07:28:21 +0000 UTC" firstStartedPulling="2025-10-02 07:28:21.842121852 +0000 UTC m=+722.874068129" lastFinishedPulling="2025-10-02 07:28:24.875892486 +0000 UTC m=+725.907838773" observedRunningTime="2025-10-02 07:28:25.172408116 +0000 UTC m=+726.204354413" watchObservedRunningTime="2025-10-02 07:28:28.192883547 +0000 UTC m=+729.224829834" Oct 02 07:28:28 crc kubenswrapper[4960]: I1002 07:28:28.193057 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" podStartSLOduration=2.377874958 podStartE2EDuration="7.193053831s" podCreationTimestamp="2025-10-02 07:28:21 +0000 UTC" firstStartedPulling="2025-10-02 07:28:22.247274788 +0000 UTC m=+723.279221075" lastFinishedPulling="2025-10-02 07:28:27.062453661 +0000 UTC m=+728.094399948" observedRunningTime="2025-10-02 07:28:28.190593547 +0000 UTC m=+729.222539834" watchObservedRunningTime="2025-10-02 07:28:28.193053831 +0000 UTC m=+729.225000118" Oct 02 07:28:29 crc kubenswrapper[4960]: I1002 07:28:29.150190 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:28:29 crc kubenswrapper[4960]: I1002 07:28:29.150710 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:28:41 crc kubenswrapper[4960]: I1002 07:28:41.949646 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-db8949b7d-zx4b9" Oct 02 07:28:43 crc kubenswrapper[4960]: I1002 07:28:43.650647 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lddcs"] Oct 02 07:28:43 crc kubenswrapper[4960]: I1002 07:28:43.651149 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" podUID="a4cf6b9e-f466-4d58-8ffd-85487c7a2531" containerName="controller-manager" containerID="cri-o://8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054" gracePeriod=30 Oct 02 07:28:43 crc kubenswrapper[4960]: I1002 07:28:43.743874 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n"] Oct 02 07:28:43 crc kubenswrapper[4960]: I1002 07:28:43.744112 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" podUID="e36d848d-519b-4e0f-bbc6-18fd6620ef2e" containerName="route-controller-manager" containerID="cri-o://a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d" gracePeriod=30 Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.126862 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.135409 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.198144 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdmrn\" (UniqueName: \"kubernetes.io/projected/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-kube-api-access-cdmrn\") pod \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.198576 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-serving-cert\") pod \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.198672 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-client-ca\") pod \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.198758 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-serving-cert\") pod \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.198853 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp498\" (UniqueName: \"kubernetes.io/projected/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-kube-api-access-tp498\") pod \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.198940 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-client-ca\") pod \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.199082 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-proxy-ca-bundles\") pod \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.199200 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-config\") pod \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\" (UID: \"a4cf6b9e-f466-4d58-8ffd-85487c7a2531\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.199303 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-config\") pod \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\" (UID: \"e36d848d-519b-4e0f-bbc6-18fd6620ef2e\") " Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.201484 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-config" (OuterVolumeSpecName: "config") pod "e36d848d-519b-4e0f-bbc6-18fd6620ef2e" (UID: "e36d848d-519b-4e0f-bbc6-18fd6620ef2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.201476 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-client-ca" (OuterVolumeSpecName: "client-ca") pod "e36d848d-519b-4e0f-bbc6-18fd6620ef2e" (UID: "e36d848d-519b-4e0f-bbc6-18fd6620ef2e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.201608 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a4cf6b9e-f466-4d58-8ffd-85487c7a2531" (UID: "a4cf6b9e-f466-4d58-8ffd-85487c7a2531"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.201855 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-client-ca" (OuterVolumeSpecName: "client-ca") pod "a4cf6b9e-f466-4d58-8ffd-85487c7a2531" (UID: "a4cf6b9e-f466-4d58-8ffd-85487c7a2531"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.203232 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-config" (OuterVolumeSpecName: "config") pod "a4cf6b9e-f466-4d58-8ffd-85487c7a2531" (UID: "a4cf6b9e-f466-4d58-8ffd-85487c7a2531"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.207393 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a4cf6b9e-f466-4d58-8ffd-85487c7a2531" (UID: "a4cf6b9e-f466-4d58-8ffd-85487c7a2531"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.208160 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-kube-api-access-cdmrn" (OuterVolumeSpecName: "kube-api-access-cdmrn") pod "e36d848d-519b-4e0f-bbc6-18fd6620ef2e" (UID: "e36d848d-519b-4e0f-bbc6-18fd6620ef2e"). InnerVolumeSpecName "kube-api-access-cdmrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.208336 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e36d848d-519b-4e0f-bbc6-18fd6620ef2e" (UID: "e36d848d-519b-4e0f-bbc6-18fd6620ef2e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.218727 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-kube-api-access-tp498" (OuterVolumeSpecName: "kube-api-access-tp498") pod "a4cf6b9e-f466-4d58-8ffd-85487c7a2531" (UID: "a4cf6b9e-f466-4d58-8ffd-85487c7a2531"). InnerVolumeSpecName "kube-api-access-tp498". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.279678 4960 generic.go:334] "Generic (PLEG): container finished" podID="e36d848d-519b-4e0f-bbc6-18fd6620ef2e" containerID="a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d" exitCode=0 Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.279799 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" event={"ID":"e36d848d-519b-4e0f-bbc6-18fd6620ef2e","Type":"ContainerDied","Data":"a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d"} Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.279897 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" event={"ID":"e36d848d-519b-4e0f-bbc6-18fd6620ef2e","Type":"ContainerDied","Data":"bac05e546f09dd3315a11a2caaa608167331f17695309113be249122aa9f65ed"} Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.279922 4960 scope.go:117] "RemoveContainer" containerID="a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.280166 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.281687 4960 generic.go:334] "Generic (PLEG): container finished" podID="a4cf6b9e-f466-4d58-8ffd-85487c7a2531" containerID="8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054" exitCode=0 Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.281729 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" event={"ID":"a4cf6b9e-f466-4d58-8ffd-85487c7a2531","Type":"ContainerDied","Data":"8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054"} Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.281753 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" event={"ID":"a4cf6b9e-f466-4d58-8ffd-85487c7a2531","Type":"ContainerDied","Data":"ba1f211115ee5780bf36e05062363db4c2812fe062fea045039ea992dc12fe95"} Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.281831 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lddcs" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301081 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301542 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301611 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301673 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdmrn\" (UniqueName: \"kubernetes.io/projected/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-kube-api-access-cdmrn\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301742 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301795 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301845 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e36d848d-519b-4e0f-bbc6-18fd6620ef2e-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301894 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp498\" (UniqueName: \"kubernetes.io/projected/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-kube-api-access-tp498\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.301947 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4cf6b9e-f466-4d58-8ffd-85487c7a2531-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.303194 4960 scope.go:117] "RemoveContainer" containerID="a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d" Oct 02 07:28:44 crc kubenswrapper[4960]: E1002 07:28:44.303823 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d\": container with ID starting with a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d not found: ID does not exist" containerID="a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.303866 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d"} err="failed to get container status \"a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d\": rpc error: code = NotFound desc = could not find container \"a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d\": container with ID starting with a4e0422eba039ef2a9579c8111ab95a97bfcedbdf00e8db38f453cae4d716d8d not found: ID does not exist" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.303894 4960 scope.go:117] "RemoveContainer" containerID="8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.321372 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n"] Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.322929 4960 scope.go:117] "RemoveContainer" containerID="8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.327185 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vcb9n"] Oct 02 07:28:44 crc kubenswrapper[4960]: E1002 07:28:44.327579 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054\": container with ID starting with 8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054 not found: ID does not exist" containerID="8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.327723 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054"} err="failed to get container status \"8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054\": rpc error: code = NotFound desc = could not find container \"8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054\": container with ID starting with 8b121e432644dcf0fe7c23edecc54b8af272d12f8641da8c9ead4eb1001cc054 not found: ID does not exist" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.341039 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e36d848d-519b-4e0f-bbc6-18fd6620ef2e" path="/var/lib/kubelet/pods/e36d848d-519b-4e0f-bbc6-18fd6620ef2e/volumes" Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.341591 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lddcs"] Oct 02 07:28:44 crc kubenswrapper[4960]: I1002 07:28:44.341627 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lddcs"] Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.281554 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g"] Oct 02 07:28:45 crc kubenswrapper[4960]: E1002 07:28:45.282006 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36d848d-519b-4e0f-bbc6-18fd6620ef2e" containerName="route-controller-manager" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.282060 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36d848d-519b-4e0f-bbc6-18fd6620ef2e" containerName="route-controller-manager" Oct 02 07:28:45 crc kubenswrapper[4960]: E1002 07:28:45.282109 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4cf6b9e-f466-4d58-8ffd-85487c7a2531" containerName="controller-manager" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.282123 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4cf6b9e-f466-4d58-8ffd-85487c7a2531" containerName="controller-manager" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.282363 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4cf6b9e-f466-4d58-8ffd-85487c7a2531" containerName="controller-manager" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.282384 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e36d848d-519b-4e0f-bbc6-18fd6620ef2e" containerName="route-controller-manager" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.283191 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.302054 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.302081 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.302085 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.302061 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.302056 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.306060 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g"] Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.306697 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 07:28:45 crc kubenswrapper[4960]: E1002 07:28:45.310707 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-cmf6b serving-cert], unattached volumes=[], failed to process volumes=[client-ca config kube-api-access-cmf6b serving-cert]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" podUID="121e5736-3b71-41f8-a441-b645842d7c27" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.314896 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6749cd675-rmm2s"] Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.315819 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.317479 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.318100 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.318114 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.318280 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.321230 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.326442 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.326533 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.330513 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g"] Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.333563 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6749cd675-rmm2s"] Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.418145 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-config\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.418765 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/121e5736-3b71-41f8-a441-b645842d7c27-serving-cert\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.418813 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw4nf\" (UniqueName: \"kubernetes.io/projected/55cf36af-481a-4592-adbd-f6924a8e7d23-kube-api-access-lw4nf\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.418844 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-config\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.418892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmf6b\" (UniqueName: \"kubernetes.io/projected/121e5736-3b71-41f8-a441-b645842d7c27-kube-api-access-cmf6b\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.419156 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-client-ca\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.419223 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-proxy-ca-bundles\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.419291 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55cf36af-481a-4592-adbd-f6924a8e7d23-serving-cert\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.419532 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-client-ca\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521193 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-client-ca\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521246 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-config\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521290 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/121e5736-3b71-41f8-a441-b645842d7c27-serving-cert\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521318 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw4nf\" (UniqueName: \"kubernetes.io/projected/55cf36af-481a-4592-adbd-f6924a8e7d23-kube-api-access-lw4nf\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521336 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-config\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521361 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmf6b\" (UniqueName: \"kubernetes.io/projected/121e5736-3b71-41f8-a441-b645842d7c27-kube-api-access-cmf6b\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521389 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-client-ca\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521408 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-proxy-ca-bundles\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.521425 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55cf36af-481a-4592-adbd-f6924a8e7d23-serving-cert\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.522582 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-client-ca\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.523776 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-config\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.524066 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-client-ca\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.524733 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-proxy-ca-bundles\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.525440 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55cf36af-481a-4592-adbd-f6924a8e7d23-config\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.526508 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/121e5736-3b71-41f8-a441-b645842d7c27-serving-cert\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.529665 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55cf36af-481a-4592-adbd-f6924a8e7d23-serving-cert\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.542906 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmf6b\" (UniqueName: \"kubernetes.io/projected/121e5736-3b71-41f8-a441-b645842d7c27-kube-api-access-cmf6b\") pod \"route-controller-manager-7c596b98f6-n2m7g\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.543692 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw4nf\" (UniqueName: \"kubernetes.io/projected/55cf36af-481a-4592-adbd-f6924a8e7d23-kube-api-access-lw4nf\") pod \"controller-manager-6749cd675-rmm2s\" (UID: \"55cf36af-481a-4592-adbd-f6924a8e7d23\") " pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:45 crc kubenswrapper[4960]: I1002 07:28:45.669814 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.031808 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6749cd675-rmm2s"] Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.300120 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.301611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" event={"ID":"55cf36af-481a-4592-adbd-f6924a8e7d23","Type":"ContainerStarted","Data":"db0732383cf823935ec0e1f6f3ac12cdcb066af4005bbd3bd2da6ad9dcf0e81d"} Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.301639 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.301649 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" event={"ID":"55cf36af-481a-4592-adbd-f6924a8e7d23","Type":"ContainerStarted","Data":"b8b1c13a578305a1f1145d14191484f6a15905dacf972f586d451756992dfdd6"} Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.313633 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.315271 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.326784 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6749cd675-rmm2s" podStartSLOduration=1.326766218 podStartE2EDuration="1.326766218s" podCreationTimestamp="2025-10-02 07:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:28:46.322303909 +0000 UTC m=+747.354250206" watchObservedRunningTime="2025-10-02 07:28:46.326766218 +0000 UTC m=+747.358712515" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.339151 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4cf6b9e-f466-4d58-8ffd-85487c7a2531" path="/var/lib/kubelet/pods/a4cf6b9e-f466-4d58-8ffd-85487c7a2531/volumes" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.434557 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-config\") pod \"121e5736-3b71-41f8-a441-b645842d7c27\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.435034 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmf6b\" (UniqueName: \"kubernetes.io/projected/121e5736-3b71-41f8-a441-b645842d7c27-kube-api-access-cmf6b\") pod \"121e5736-3b71-41f8-a441-b645842d7c27\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.435197 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/121e5736-3b71-41f8-a441-b645842d7c27-serving-cert\") pod \"121e5736-3b71-41f8-a441-b645842d7c27\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.435364 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-client-ca\") pod \"121e5736-3b71-41f8-a441-b645842d7c27\" (UID: \"121e5736-3b71-41f8-a441-b645842d7c27\") " Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.435511 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-config" (OuterVolumeSpecName: "config") pod "121e5736-3b71-41f8-a441-b645842d7c27" (UID: "121e5736-3b71-41f8-a441-b645842d7c27"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.435920 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.436022 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-client-ca" (OuterVolumeSpecName: "client-ca") pod "121e5736-3b71-41f8-a441-b645842d7c27" (UID: "121e5736-3b71-41f8-a441-b645842d7c27"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.443177 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/121e5736-3b71-41f8-a441-b645842d7c27-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "121e5736-3b71-41f8-a441-b645842d7c27" (UID: "121e5736-3b71-41f8-a441-b645842d7c27"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.443309 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/121e5736-3b71-41f8-a441-b645842d7c27-kube-api-access-cmf6b" (OuterVolumeSpecName: "kube-api-access-cmf6b") pod "121e5736-3b71-41f8-a441-b645842d7c27" (UID: "121e5736-3b71-41f8-a441-b645842d7c27"). InnerVolumeSpecName "kube-api-access-cmf6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.537260 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/121e5736-3b71-41f8-a441-b645842d7c27-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.537301 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmf6b\" (UniqueName: \"kubernetes.io/projected/121e5736-3b71-41f8-a441-b645842d7c27-kube-api-access-cmf6b\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:46 crc kubenswrapper[4960]: I1002 07:28:46.537315 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/121e5736-3b71-41f8-a441-b645842d7c27-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.305503 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.344324 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g"] Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.348144 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q"] Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.349070 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.351671 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c596b98f6-n2m7g"] Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.351879 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.352117 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.352138 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.353332 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.353467 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.353640 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.371088 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q"] Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.453788 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n65hw\" (UniqueName: \"kubernetes.io/projected/99b159f4-2c11-42e4-9bd5-3195e167d821-kube-api-access-n65hw\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.453847 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99b159f4-2c11-42e4-9bd5-3195e167d821-client-ca\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.453922 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b159f4-2c11-42e4-9bd5-3195e167d821-config\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.453941 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99b159f4-2c11-42e4-9bd5-3195e167d821-serving-cert\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.555179 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b159f4-2c11-42e4-9bd5-3195e167d821-config\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.555237 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99b159f4-2c11-42e4-9bd5-3195e167d821-serving-cert\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.555304 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n65hw\" (UniqueName: \"kubernetes.io/projected/99b159f4-2c11-42e4-9bd5-3195e167d821-kube-api-access-n65hw\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.555345 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99b159f4-2c11-42e4-9bd5-3195e167d821-client-ca\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.556446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99b159f4-2c11-42e4-9bd5-3195e167d821-config\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.556468 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99b159f4-2c11-42e4-9bd5-3195e167d821-client-ca\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.570083 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99b159f4-2c11-42e4-9bd5-3195e167d821-serving-cert\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.586696 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n65hw\" (UniqueName: \"kubernetes.io/projected/99b159f4-2c11-42e4-9bd5-3195e167d821-kube-api-access-n65hw\") pod \"route-controller-manager-6cddc6c9dc-q442q\" (UID: \"99b159f4-2c11-42e4-9bd5-3195e167d821\") " pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:47 crc kubenswrapper[4960]: I1002 07:28:47.667880 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:48 crc kubenswrapper[4960]: I1002 07:28:48.174678 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q"] Oct 02 07:28:48 crc kubenswrapper[4960]: W1002 07:28:48.179083 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99b159f4_2c11_42e4_9bd5_3195e167d821.slice/crio-9dcb51d21bfb4b1e0d113cb675ea4c3ff3d5976fcc3bcf64ad561f1cd786402b WatchSource:0}: Error finding container 9dcb51d21bfb4b1e0d113cb675ea4c3ff3d5976fcc3bcf64ad561f1cd786402b: Status 404 returned error can't find the container with id 9dcb51d21bfb4b1e0d113cb675ea4c3ff3d5976fcc3bcf64ad561f1cd786402b Oct 02 07:28:48 crc kubenswrapper[4960]: I1002 07:28:48.315588 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" event={"ID":"99b159f4-2c11-42e4-9bd5-3195e167d821","Type":"ContainerStarted","Data":"06ac002fded4eff31475463a01d22dae8f0c934e416e83b288cb2b742d684e4d"} Oct 02 07:28:48 crc kubenswrapper[4960]: I1002 07:28:48.316184 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" event={"ID":"99b159f4-2c11-42e4-9bd5-3195e167d821","Type":"ContainerStarted","Data":"9dcb51d21bfb4b1e0d113cb675ea4c3ff3d5976fcc3bcf64ad561f1cd786402b"} Oct 02 07:28:48 crc kubenswrapper[4960]: I1002 07:28:48.343064 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="121e5736-3b71-41f8-a441-b645842d7c27" path="/var/lib/kubelet/pods/121e5736-3b71-41f8-a441-b645842d7c27/volumes" Oct 02 07:28:49 crc kubenswrapper[4960]: I1002 07:28:49.322012 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:49 crc kubenswrapper[4960]: I1002 07:28:49.331280 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" Oct 02 07:28:49 crc kubenswrapper[4960]: I1002 07:28:49.343956 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6cddc6c9dc-q442q" podStartSLOduration=4.343936112 podStartE2EDuration="4.343936112s" podCreationTimestamp="2025-10-02 07:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:28:49.343212032 +0000 UTC m=+750.375158329" watchObservedRunningTime="2025-10-02 07:28:49.343936112 +0000 UTC m=+750.375882399" Oct 02 07:28:55 crc kubenswrapper[4960]: I1002 07:28:55.304105 4960 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 07:28:59 crc kubenswrapper[4960]: I1002 07:28:59.150254 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:28:59 crc kubenswrapper[4960]: I1002 07:28:59.150940 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:29:01 crc kubenswrapper[4960]: I1002 07:29:01.551662 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-cf55b8d8b-rzqpg" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.340879 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-tr9mf"] Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.344228 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.347102 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-f94cb" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.347481 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.347821 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.351943 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx"] Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.352897 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.355832 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.379198 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx"] Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397264 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-reloader\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397333 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-sockets\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397363 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-metrics\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397386 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-startup\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397438 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8f2f\" (UniqueName: \"kubernetes.io/projected/ab470676-e7e9-40a7-96db-3b1cc494dc0a-kube-api-access-q8f2f\") pod \"frr-k8s-webhook-server-64bf5d555-68xcx\" (UID: \"ab470676-e7e9-40a7-96db-3b1cc494dc0a\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397479 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-metrics-certs\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397551 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf82s\" (UniqueName: \"kubernetes.io/projected/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-kube-api-access-lf82s\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397599 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-conf\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.397640 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab470676-e7e9-40a7-96db-3b1cc494dc0a-cert\") pod \"frr-k8s-webhook-server-64bf5d555-68xcx\" (UID: \"ab470676-e7e9-40a7-96db-3b1cc494dc0a\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499277 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-conf\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499358 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab470676-e7e9-40a7-96db-3b1cc494dc0a-cert\") pod \"frr-k8s-webhook-server-64bf5d555-68xcx\" (UID: \"ab470676-e7e9-40a7-96db-3b1cc494dc0a\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-reloader\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499406 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-sockets\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499429 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-metrics\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499453 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-startup\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499490 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8f2f\" (UniqueName: \"kubernetes.io/projected/ab470676-e7e9-40a7-96db-3b1cc494dc0a-kube-api-access-q8f2f\") pod \"frr-k8s-webhook-server-64bf5d555-68xcx\" (UID: \"ab470676-e7e9-40a7-96db-3b1cc494dc0a\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499523 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-metrics-certs\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.499562 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf82s\" (UniqueName: \"kubernetes.io/projected/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-kube-api-access-lf82s\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.500561 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-conf\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.500664 4960 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.500735 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab470676-e7e9-40a7-96db-3b1cc494dc0a-cert podName:ab470676-e7e9-40a7-96db-3b1cc494dc0a nodeName:}" failed. No retries permitted until 2025-10-02 07:29:03.000711442 +0000 UTC m=+764.032657729 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ab470676-e7e9-40a7-96db-3b1cc494dc0a-cert") pod "frr-k8s-webhook-server-64bf5d555-68xcx" (UID: "ab470676-e7e9-40a7-96db-3b1cc494dc0a") : secret "frr-k8s-webhook-server-cert" not found Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.500576 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-metrics\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.501172 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-reloader\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.501546 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-startup\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.501714 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-frr-sockets\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.502985 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5j29v"] Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.504577 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.508172 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-sdmzg" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.508355 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.515830 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-metrics-certs\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.533285 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.534077 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.539707 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf82s\" (UniqueName: \"kubernetes.io/projected/ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0-kube-api-access-lf82s\") pod \"frr-k8s-tr9mf\" (UID: \"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0\") " pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.539715 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-rd2k2"] Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.541029 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.545374 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.580614 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-rd2k2"] Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.585754 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8f2f\" (UniqueName: \"kubernetes.io/projected/ab470676-e7e9-40a7-96db-3b1cc494dc0a-kube-api-access-q8f2f\") pod \"frr-k8s-webhook-server-64bf5d555-68xcx\" (UID: \"ab470676-e7e9-40a7-96db-3b1cc494dc0a\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.603717 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.603779 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-metrics-certs\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.603801 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-metrics-certs\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.603833 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-cert\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.603868 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg49h\" (UniqueName: \"kubernetes.io/projected/2d48cc2b-907f-4811-8cb0-779e2799638d-kube-api-access-xg49h\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.603923 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a4cae542-7390-4af8-8c12-97addccc5b4b-metallb-excludel2\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.603941 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l67xl\" (UniqueName: \"kubernetes.io/projected/a4cae542-7390-4af8-8c12-97addccc5b4b-kube-api-access-l67xl\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.671756 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.705409 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a4cae542-7390-4af8-8c12-97addccc5b4b-metallb-excludel2\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.705449 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l67xl\" (UniqueName: \"kubernetes.io/projected/a4cae542-7390-4af8-8c12-97addccc5b4b-kube-api-access-l67xl\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.705484 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.705518 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-metrics-certs\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.705534 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-metrics-certs\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.705550 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-cert\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.705581 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg49h\" (UniqueName: \"kubernetes.io/projected/2d48cc2b-907f-4811-8cb0-779e2799638d-kube-api-access-xg49h\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.706092 4960 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.706144 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist podName:a4cae542-7390-4af8-8c12-97addccc5b4b nodeName:}" failed. No retries permitted until 2025-10-02 07:29:03.20612812 +0000 UTC m=+764.238074407 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist") pod "speaker-5j29v" (UID: "a4cae542-7390-4af8-8c12-97addccc5b4b") : secret "metallb-memberlist" not found Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.706217 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a4cae542-7390-4af8-8c12-97addccc5b4b-metallb-excludel2\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.706274 4960 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.706297 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-metrics-certs podName:a4cae542-7390-4af8-8c12-97addccc5b4b nodeName:}" failed. No retries permitted until 2025-10-02 07:29:03.206289864 +0000 UTC m=+764.238236151 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-metrics-certs") pod "speaker-5j29v" (UID: "a4cae542-7390-4af8-8c12-97addccc5b4b") : secret "speaker-certs-secret" not found Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.706334 4960 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 02 07:29:02 crc kubenswrapper[4960]: E1002 07:29:02.706352 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-metrics-certs podName:2d48cc2b-907f-4811-8cb0-779e2799638d nodeName:}" failed. No retries permitted until 2025-10-02 07:29:03.206345186 +0000 UTC m=+764.238291473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-metrics-certs") pod "controller-68d546b9d8-rd2k2" (UID: "2d48cc2b-907f-4811-8cb0-779e2799638d") : secret "controller-certs-secret" not found Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.711303 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-cert\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.729618 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg49h\" (UniqueName: \"kubernetes.io/projected/2d48cc2b-907f-4811-8cb0-779e2799638d-kube-api-access-xg49h\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:02 crc kubenswrapper[4960]: I1002 07:29:02.745495 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l67xl\" (UniqueName: \"kubernetes.io/projected/a4cae542-7390-4af8-8c12-97addccc5b4b-kube-api-access-l67xl\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.011351 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab470676-e7e9-40a7-96db-3b1cc494dc0a-cert\") pod \"frr-k8s-webhook-server-64bf5d555-68xcx\" (UID: \"ab470676-e7e9-40a7-96db-3b1cc494dc0a\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.017398 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab470676-e7e9-40a7-96db-3b1cc494dc0a-cert\") pod \"frr-k8s-webhook-server-64bf5d555-68xcx\" (UID: \"ab470676-e7e9-40a7-96db-3b1cc494dc0a\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.214348 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.214436 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-metrics-certs\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.214465 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-metrics-certs\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:03 crc kubenswrapper[4960]: E1002 07:29:03.214600 4960 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 07:29:03 crc kubenswrapper[4960]: E1002 07:29:03.214711 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist podName:a4cae542-7390-4af8-8c12-97addccc5b4b nodeName:}" failed. No retries permitted until 2025-10-02 07:29:04.214679213 +0000 UTC m=+765.246625510 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist") pod "speaker-5j29v" (UID: "a4cae542-7390-4af8-8c12-97addccc5b4b") : secret "metallb-memberlist" not found Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.217689 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d48cc2b-907f-4811-8cb0-779e2799638d-metrics-certs\") pod \"controller-68d546b9d8-rd2k2\" (UID: \"2d48cc2b-907f-4811-8cb0-779e2799638d\") " pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.219917 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-metrics-certs\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.228018 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.280905 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.425215 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerStarted","Data":"61e1b9bd826af16eb3d0ef4e5c5c020638eb2b69ad493c6e54da42624417550b"} Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.659458 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-rd2k2"] Oct 02 07:29:03 crc kubenswrapper[4960]: I1002 07:29:03.837811 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx"] Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.233101 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.239031 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4cae542-7390-4af8-8c12-97addccc5b4b-memberlist\") pod \"speaker-5j29v\" (UID: \"a4cae542-7390-4af8-8c12-97addccc5b4b\") " pod="metallb-system/speaker-5j29v" Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.410700 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5j29v" Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.436357 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" event={"ID":"ab470676-e7e9-40a7-96db-3b1cc494dc0a","Type":"ContainerStarted","Data":"fde97e84923a87c9ceb68584cd8acfc30b75dbaba61b25357d1c7aaa7ab62dd9"} Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.440661 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-rd2k2" event={"ID":"2d48cc2b-907f-4811-8cb0-779e2799638d","Type":"ContainerStarted","Data":"ef9b916460896da559baf49a6488ed3e02d70682547b7fad71eecd93c3a6e34d"} Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.440704 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-rd2k2" event={"ID":"2d48cc2b-907f-4811-8cb0-779e2799638d","Type":"ContainerStarted","Data":"07dcfb6d8d10ce77fff6a91734b4414671f91788a0a9ba56f513acc7095ca799"} Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.440719 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-rd2k2" event={"ID":"2d48cc2b-907f-4811-8cb0-779e2799638d","Type":"ContainerStarted","Data":"55f8ccffc2dc5b31f5d16531e5f44096ad7e86599e706b73de864e7a1c402823"} Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.440798 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:04 crc kubenswrapper[4960]: I1002 07:29:04.460299 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-rd2k2" podStartSLOduration=2.460282428 podStartE2EDuration="2.460282428s" podCreationTimestamp="2025-10-02 07:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:29:04.459148078 +0000 UTC m=+765.491094365" watchObservedRunningTime="2025-10-02 07:29:04.460282428 +0000 UTC m=+765.492228715" Oct 02 07:29:05 crc kubenswrapper[4960]: I1002 07:29:05.453017 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5j29v" event={"ID":"a4cae542-7390-4af8-8c12-97addccc5b4b","Type":"ContainerStarted","Data":"71c7d9830b2d2a5b72bdcbca0d8bb9a127547d5818028860c3fcfe7d50221f91"} Oct 02 07:29:05 crc kubenswrapper[4960]: I1002 07:29:05.453067 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5j29v" event={"ID":"a4cae542-7390-4af8-8c12-97addccc5b4b","Type":"ContainerStarted","Data":"f696e8d0f62ef85e5840d36008ea0ffcb2a1a11360d9feee9b7b6c29ff1ed5c4"} Oct 02 07:29:05 crc kubenswrapper[4960]: I1002 07:29:05.453079 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5j29v" event={"ID":"a4cae542-7390-4af8-8c12-97addccc5b4b","Type":"ContainerStarted","Data":"69b69971cd007732fabb0f9763d82dc4dfd6e83d4cb42fa7cc9efa991389e562"} Oct 02 07:29:05 crc kubenswrapper[4960]: I1002 07:29:05.454298 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5j29v" Oct 02 07:29:05 crc kubenswrapper[4960]: I1002 07:29:05.476350 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5j29v" podStartSLOduration=3.476325756 podStartE2EDuration="3.476325756s" podCreationTimestamp="2025-10-02 07:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:29:05.47382426 +0000 UTC m=+766.505770547" watchObservedRunningTime="2025-10-02 07:29:05.476325756 +0000 UTC m=+766.508272043" Oct 02 07:29:11 crc kubenswrapper[4960]: I1002 07:29:11.496613 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" event={"ID":"ab470676-e7e9-40a7-96db-3b1cc494dc0a","Type":"ContainerStarted","Data":"b5a0555169e9f61a4f788b95e3bba2727c452c5e3cb9f58e3a1f7070ecb3d847"} Oct 02 07:29:11 crc kubenswrapper[4960]: I1002 07:29:11.497261 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:11 crc kubenswrapper[4960]: I1002 07:29:11.500693 4960 generic.go:334] "Generic (PLEG): container finished" podID="ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0" containerID="094e1d76095dbc3436907df2ca1295e99f015a49d8051024f8a246be76144b49" exitCode=0 Oct 02 07:29:11 crc kubenswrapper[4960]: I1002 07:29:11.500761 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerDied","Data":"094e1d76095dbc3436907df2ca1295e99f015a49d8051024f8a246be76144b49"} Oct 02 07:29:11 crc kubenswrapper[4960]: I1002 07:29:11.529312 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" podStartSLOduration=2.459014604 podStartE2EDuration="9.529264236s" podCreationTimestamp="2025-10-02 07:29:02 +0000 UTC" firstStartedPulling="2025-10-02 07:29:03.852104097 +0000 UTC m=+764.884050414" lastFinishedPulling="2025-10-02 07:29:10.922353749 +0000 UTC m=+771.954300046" observedRunningTime="2025-10-02 07:29:11.52189463 +0000 UTC m=+772.553840917" watchObservedRunningTime="2025-10-02 07:29:11.529264236 +0000 UTC m=+772.561210543" Oct 02 07:29:12 crc kubenswrapper[4960]: I1002 07:29:12.512036 4960 generic.go:334] "Generic (PLEG): container finished" podID="ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0" containerID="677f5e47d493f881e68690f8cd819688d47eba6b9b3d3ea0892d7294f1fe4ec1" exitCode=0 Oct 02 07:29:12 crc kubenswrapper[4960]: I1002 07:29:12.512189 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerDied","Data":"677f5e47d493f881e68690f8cd819688d47eba6b9b3d3ea0892d7294f1fe4ec1"} Oct 02 07:29:13 crc kubenswrapper[4960]: I1002 07:29:13.231806 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-rd2k2" Oct 02 07:29:13 crc kubenswrapper[4960]: I1002 07:29:13.522915 4960 generic.go:334] "Generic (PLEG): container finished" podID="ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0" containerID="3dc4f58a22c7526e2943c4f08d0a7bc4dc14c07fa903eddd5045ae983a38b240" exitCode=0 Oct 02 07:29:13 crc kubenswrapper[4960]: I1002 07:29:13.523002 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerDied","Data":"3dc4f58a22c7526e2943c4f08d0a7bc4dc14c07fa903eddd5045ae983a38b240"} Oct 02 07:29:14 crc kubenswrapper[4960]: I1002 07:29:14.415116 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5j29v" Oct 02 07:29:14 crc kubenswrapper[4960]: I1002 07:29:14.536616 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerStarted","Data":"21353f26eeb855adec08777e532f7f46e03070c3f0d65477202af67b8d3bcce9"} Oct 02 07:29:14 crc kubenswrapper[4960]: I1002 07:29:14.537473 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerStarted","Data":"5e0c11bc319a5dec44b42f25aeaef3bd6ecfc1a41632950db6b5488c8c4fd741"} Oct 02 07:29:14 crc kubenswrapper[4960]: I1002 07:29:14.537492 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerStarted","Data":"bd5f6a928ee68f38aad013226366bf5d696e1c36538e302fdc7894e9f246be04"} Oct 02 07:29:14 crc kubenswrapper[4960]: I1002 07:29:14.537501 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerStarted","Data":"ddb39c77b38e45a3d9fdbd3f9629b7f62d71ac4c2c9f0b3fbb25f39c8f398b15"} Oct 02 07:29:14 crc kubenswrapper[4960]: I1002 07:29:14.537510 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerStarted","Data":"48d3bae15bd99504ccfce2d2fd78bd9707974c64a40f2c99b79e83a3f29e110c"} Oct 02 07:29:15 crc kubenswrapper[4960]: I1002 07:29:15.549864 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tr9mf" event={"ID":"ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0","Type":"ContainerStarted","Data":"de2622d112148edda02fb77d491b3509504d85c4f4d3fc98b3851ef6c5176def"} Oct 02 07:29:15 crc kubenswrapper[4960]: I1002 07:29:15.550469 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:15 crc kubenswrapper[4960]: I1002 07:29:15.586395 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-tr9mf" podStartSLOduration=5.507699823 podStartE2EDuration="13.58636592s" podCreationTimestamp="2025-10-02 07:29:02 +0000 UTC" firstStartedPulling="2025-10-02 07:29:02.815557357 +0000 UTC m=+763.847503644" lastFinishedPulling="2025-10-02 07:29:10.894223444 +0000 UTC m=+771.926169741" observedRunningTime="2025-10-02 07:29:15.585628351 +0000 UTC m=+776.617574678" watchObservedRunningTime="2025-10-02 07:29:15.58636592 +0000 UTC m=+776.618312247" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.560788 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-l9cv8"] Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.562607 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l9cv8" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.565686 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.565724 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.583782 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-l9cv8"] Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.663931 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv979\" (UniqueName: \"kubernetes.io/projected/fbe1a009-3abe-4396-86c4-2afd1ec1af4a-kube-api-access-hv979\") pod \"openstack-operator-index-l9cv8\" (UID: \"fbe1a009-3abe-4396-86c4-2afd1ec1af4a\") " pod="openstack-operators/openstack-operator-index-l9cv8" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.672377 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.747676 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.765640 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv979\" (UniqueName: \"kubernetes.io/projected/fbe1a009-3abe-4396-86c4-2afd1ec1af4a-kube-api-access-hv979\") pod \"openstack-operator-index-l9cv8\" (UID: \"fbe1a009-3abe-4396-86c4-2afd1ec1af4a\") " pod="openstack-operators/openstack-operator-index-l9cv8" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.802099 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv979\" (UniqueName: \"kubernetes.io/projected/fbe1a009-3abe-4396-86c4-2afd1ec1af4a-kube-api-access-hv979\") pod \"openstack-operator-index-l9cv8\" (UID: \"fbe1a009-3abe-4396-86c4-2afd1ec1af4a\") " pod="openstack-operators/openstack-operator-index-l9cv8" Oct 02 07:29:17 crc kubenswrapper[4960]: I1002 07:29:17.889177 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l9cv8" Oct 02 07:29:18 crc kubenswrapper[4960]: I1002 07:29:18.322623 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-l9cv8"] Oct 02 07:29:18 crc kubenswrapper[4960]: I1002 07:29:18.571122 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l9cv8" event={"ID":"fbe1a009-3abe-4396-86c4-2afd1ec1af4a","Type":"ContainerStarted","Data":"37c172557f522cacb0d5a220790303c30d86d6e4ab9b4879f75dd201c9aafa94"} Oct 02 07:29:19 crc kubenswrapper[4960]: I1002 07:29:19.581604 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l9cv8" event={"ID":"fbe1a009-3abe-4396-86c4-2afd1ec1af4a","Type":"ContainerStarted","Data":"0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59"} Oct 02 07:29:19 crc kubenswrapper[4960]: I1002 07:29:19.606690 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-l9cv8" podStartSLOduration=1.7437966010000001 podStartE2EDuration="2.606665749s" podCreationTimestamp="2025-10-02 07:29:17 +0000 UTC" firstStartedPulling="2025-10-02 07:29:18.331623472 +0000 UTC m=+779.363569769" lastFinishedPulling="2025-10-02 07:29:19.19449263 +0000 UTC m=+780.226438917" observedRunningTime="2025-10-02 07:29:19.605932329 +0000 UTC m=+780.637878656" watchObservedRunningTime="2025-10-02 07:29:19.606665749 +0000 UTC m=+780.638612046" Oct 02 07:29:20 crc kubenswrapper[4960]: I1002 07:29:20.931938 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-l9cv8"] Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.531802 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-htqft"] Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.532654 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.535608 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-hnq2f" Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.541566 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-htqft"] Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.592145 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-l9cv8" podUID="fbe1a009-3abe-4396-86c4-2afd1ec1af4a" containerName="registry-server" containerID="cri-o://0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59" gracePeriod=2 Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.635210 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m92g6\" (UniqueName: \"kubernetes.io/projected/12f72b6d-7325-428c-8a30-fd0d75158b31-kube-api-access-m92g6\") pod \"openstack-operator-index-htqft\" (UID: \"12f72b6d-7325-428c-8a30-fd0d75158b31\") " pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.737443 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m92g6\" (UniqueName: \"kubernetes.io/projected/12f72b6d-7325-428c-8a30-fd0d75158b31-kube-api-access-m92g6\") pod \"openstack-operator-index-htqft\" (UID: \"12f72b6d-7325-428c-8a30-fd0d75158b31\") " pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.775296 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m92g6\" (UniqueName: \"kubernetes.io/projected/12f72b6d-7325-428c-8a30-fd0d75158b31-kube-api-access-m92g6\") pod \"openstack-operator-index-htqft\" (UID: \"12f72b6d-7325-428c-8a30-fd0d75158b31\") " pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:21 crc kubenswrapper[4960]: I1002 07:29:21.853424 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.040373 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l9cv8" Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.144751 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv979\" (UniqueName: \"kubernetes.io/projected/fbe1a009-3abe-4396-86c4-2afd1ec1af4a-kube-api-access-hv979\") pod \"fbe1a009-3abe-4396-86c4-2afd1ec1af4a\" (UID: \"fbe1a009-3abe-4396-86c4-2afd1ec1af4a\") " Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.149410 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbe1a009-3abe-4396-86c4-2afd1ec1af4a-kube-api-access-hv979" (OuterVolumeSpecName: "kube-api-access-hv979") pod "fbe1a009-3abe-4396-86c4-2afd1ec1af4a" (UID: "fbe1a009-3abe-4396-86c4-2afd1ec1af4a"). InnerVolumeSpecName "kube-api-access-hv979". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.246875 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv979\" (UniqueName: \"kubernetes.io/projected/fbe1a009-3abe-4396-86c4-2afd1ec1af4a-kube-api-access-hv979\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.379203 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-htqft"] Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.602131 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-htqft" event={"ID":"12f72b6d-7325-428c-8a30-fd0d75158b31","Type":"ContainerStarted","Data":"142254f0c0ed76f3bb229c407c9c987226929a069e0b3155876719dc12c1ba91"} Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.605014 4960 generic.go:334] "Generic (PLEG): container finished" podID="fbe1a009-3abe-4396-86c4-2afd1ec1af4a" containerID="0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59" exitCode=0 Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.605090 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l9cv8" event={"ID":"fbe1a009-3abe-4396-86c4-2afd1ec1af4a","Type":"ContainerDied","Data":"0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59"} Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.605118 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l9cv8" event={"ID":"fbe1a009-3abe-4396-86c4-2afd1ec1af4a","Type":"ContainerDied","Data":"37c172557f522cacb0d5a220790303c30d86d6e4ab9b4879f75dd201c9aafa94"} Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.605160 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l9cv8" Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.605172 4960 scope.go:117] "RemoveContainer" containerID="0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59" Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.631371 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-l9cv8"] Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.635199 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-l9cv8"] Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.642719 4960 scope.go:117] "RemoveContainer" containerID="0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59" Oct 02 07:29:22 crc kubenswrapper[4960]: E1002 07:29:22.643361 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59\": container with ID starting with 0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59 not found: ID does not exist" containerID="0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59" Oct 02 07:29:22 crc kubenswrapper[4960]: I1002 07:29:22.643435 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59"} err="failed to get container status \"0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59\": rpc error: code = NotFound desc = could not find container \"0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59\": container with ID starting with 0bb8df6c42188efd65d4a30f33978bc3c881814566aa86ae02bb580afcca9a59 not found: ID does not exist" Oct 02 07:29:23 crc kubenswrapper[4960]: I1002 07:29:23.287503 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-68xcx" Oct 02 07:29:23 crc kubenswrapper[4960]: I1002 07:29:23.613898 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-htqft" event={"ID":"12f72b6d-7325-428c-8a30-fd0d75158b31","Type":"ContainerStarted","Data":"000e943edfc0b52fce0c4b80fd7cdb35ff1db31f831b9e20339e054a435945c5"} Oct 02 07:29:23 crc kubenswrapper[4960]: I1002 07:29:23.637328 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-htqft" podStartSLOduration=2.192318326 podStartE2EDuration="2.637303683s" podCreationTimestamp="2025-10-02 07:29:21 +0000 UTC" firstStartedPulling="2025-10-02 07:29:22.384750109 +0000 UTC m=+783.416696396" lastFinishedPulling="2025-10-02 07:29:22.829735466 +0000 UTC m=+783.861681753" observedRunningTime="2025-10-02 07:29:23.637050687 +0000 UTC m=+784.668997014" watchObservedRunningTime="2025-10-02 07:29:23.637303683 +0000 UTC m=+784.669249960" Oct 02 07:29:24 crc kubenswrapper[4960]: I1002 07:29:24.340777 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbe1a009-3abe-4396-86c4-2afd1ec1af4a" path="/var/lib/kubelet/pods/fbe1a009-3abe-4396-86c4-2afd1ec1af4a/volumes" Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.150278 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.151325 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.151380 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.152160 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"856ec690ae49b3ae3dbe5a23fd799e7b1fd275afa9ab067d75708fa279072617"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.152238 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://856ec690ae49b3ae3dbe5a23fd799e7b1fd275afa9ab067d75708fa279072617" gracePeriod=600 Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.663755 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="856ec690ae49b3ae3dbe5a23fd799e7b1fd275afa9ab067d75708fa279072617" exitCode=0 Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.663916 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"856ec690ae49b3ae3dbe5a23fd799e7b1fd275afa9ab067d75708fa279072617"} Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.664308 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"5fc62ae8f7b8e010f4106ce66a46ce004d1ca089113eca9aa199c28f45628180"} Oct 02 07:29:29 crc kubenswrapper[4960]: I1002 07:29:29.664349 4960 scope.go:117] "RemoveContainer" containerID="078c8db152d306f847e249781287a9e53d5a972cc64796b1689df8c3d5c083f5" Oct 02 07:29:31 crc kubenswrapper[4960]: I1002 07:29:31.854551 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:31 crc kubenswrapper[4960]: I1002 07:29:31.855316 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:31 crc kubenswrapper[4960]: I1002 07:29:31.905554 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:32 crc kubenswrapper[4960]: I1002 07:29:32.676877 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-tr9mf" Oct 02 07:29:32 crc kubenswrapper[4960]: I1002 07:29:32.751359 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-htqft" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.610578 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s"] Oct 02 07:29:38 crc kubenswrapper[4960]: E1002 07:29:38.612438 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe1a009-3abe-4396-86c4-2afd1ec1af4a" containerName="registry-server" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.612474 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe1a009-3abe-4396-86c4-2afd1ec1af4a" containerName="registry-server" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.612901 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe1a009-3abe-4396-86c4-2afd1ec1af4a" containerName="registry-server" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.616288 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.619813 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-t64st" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.643162 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s"] Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.742093 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwblt\" (UniqueName: \"kubernetes.io/projected/b728cb6d-14fc-437a-abe1-836ee912d913-kube-api-access-gwblt\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.742196 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-bundle\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.742270 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-util\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.843823 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-bundle\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.844012 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-util\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.844127 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwblt\" (UniqueName: \"kubernetes.io/projected/b728cb6d-14fc-437a-abe1-836ee912d913-kube-api-access-gwblt\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.844734 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-bundle\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.844923 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-util\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.882869 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwblt\" (UniqueName: \"kubernetes.io/projected/b728cb6d-14fc-437a-abe1-836ee912d913-kube-api-access-gwblt\") pod \"0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:38 crc kubenswrapper[4960]: I1002 07:29:38.967451 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:39 crc kubenswrapper[4960]: I1002 07:29:39.496794 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s"] Oct 02 07:29:39 crc kubenswrapper[4960]: I1002 07:29:39.759132 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" event={"ID":"b728cb6d-14fc-437a-abe1-836ee912d913","Type":"ContainerStarted","Data":"53e76100a14eeb1c9d1224fac80b6fd211f5c533c4e72c99d8c647cf3bd622b5"} Oct 02 07:29:39 crc kubenswrapper[4960]: I1002 07:29:39.759632 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" event={"ID":"b728cb6d-14fc-437a-abe1-836ee912d913","Type":"ContainerStarted","Data":"3cfb27391257a6ea851c43277ad9b43561a066051e6734c3b8dba3aa7df1929b"} Oct 02 07:29:40 crc kubenswrapper[4960]: I1002 07:29:40.771332 4960 generic.go:334] "Generic (PLEG): container finished" podID="b728cb6d-14fc-437a-abe1-836ee912d913" containerID="53e76100a14eeb1c9d1224fac80b6fd211f5c533c4e72c99d8c647cf3bd622b5" exitCode=0 Oct 02 07:29:40 crc kubenswrapper[4960]: I1002 07:29:40.771397 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" event={"ID":"b728cb6d-14fc-437a-abe1-836ee912d913","Type":"ContainerDied","Data":"53e76100a14eeb1c9d1224fac80b6fd211f5c533c4e72c99d8c647cf3bd622b5"} Oct 02 07:29:41 crc kubenswrapper[4960]: I1002 07:29:41.752612 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fgj42"] Oct 02 07:29:41 crc kubenswrapper[4960]: I1002 07:29:41.759195 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:41 crc kubenswrapper[4960]: I1002 07:29:41.764714 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fgj42"] Oct 02 07:29:41 crc kubenswrapper[4960]: I1002 07:29:41.919073 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9ts8\" (UniqueName: \"kubernetes.io/projected/c7531862-540a-495a-8472-edaf3231c0b7-kube-api-access-n9ts8\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:41 crc kubenswrapper[4960]: I1002 07:29:41.919207 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-catalog-content\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:41 crc kubenswrapper[4960]: I1002 07:29:41.919345 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-utilities\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.020942 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-utilities\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.021523 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9ts8\" (UniqueName: \"kubernetes.io/projected/c7531862-540a-495a-8472-edaf3231c0b7-kube-api-access-n9ts8\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.021582 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-utilities\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.021644 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-catalog-content\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.022256 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-catalog-content\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.044700 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9ts8\" (UniqueName: \"kubernetes.io/projected/c7531862-540a-495a-8472-edaf3231c0b7-kube-api-access-n9ts8\") pod \"redhat-operators-fgj42\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.138248 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.382451 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fgj42"] Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.789594 4960 generic.go:334] "Generic (PLEG): container finished" podID="b728cb6d-14fc-437a-abe1-836ee912d913" containerID="2b5223f8c3fb328c6a9ce63392b0d2a5e8ebe4ffb5e55920c1e918d820b5b247" exitCode=0 Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.789675 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" event={"ID":"b728cb6d-14fc-437a-abe1-836ee912d913","Type":"ContainerDied","Data":"2b5223f8c3fb328c6a9ce63392b0d2a5e8ebe4ffb5e55920c1e918d820b5b247"} Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.792066 4960 generic.go:334] "Generic (PLEG): container finished" podID="c7531862-540a-495a-8472-edaf3231c0b7" containerID="d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e" exitCode=0 Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.792124 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgj42" event={"ID":"c7531862-540a-495a-8472-edaf3231c0b7","Type":"ContainerDied","Data":"d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e"} Oct 02 07:29:42 crc kubenswrapper[4960]: I1002 07:29:42.792161 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgj42" event={"ID":"c7531862-540a-495a-8472-edaf3231c0b7","Type":"ContainerStarted","Data":"ceb38bcc9b8b6a70cce0c34c1d9bbb62ab2f277a7c714d52c3c8aab2b3bfd072"} Oct 02 07:29:43 crc kubenswrapper[4960]: I1002 07:29:43.801271 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgj42" event={"ID":"c7531862-540a-495a-8472-edaf3231c0b7","Type":"ContainerStarted","Data":"2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c"} Oct 02 07:29:43 crc kubenswrapper[4960]: I1002 07:29:43.802970 4960 generic.go:334] "Generic (PLEG): container finished" podID="b728cb6d-14fc-437a-abe1-836ee912d913" containerID="82c19a454e302892fc2f8e07593f850389ecee16859b58ef4c98a556a29e67a4" exitCode=0 Oct 02 07:29:43 crc kubenswrapper[4960]: I1002 07:29:43.803028 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" event={"ID":"b728cb6d-14fc-437a-abe1-836ee912d913","Type":"ContainerDied","Data":"82c19a454e302892fc2f8e07593f850389ecee16859b58ef4c98a556a29e67a4"} Oct 02 07:29:44 crc kubenswrapper[4960]: I1002 07:29:44.814637 4960 generic.go:334] "Generic (PLEG): container finished" podID="c7531862-540a-495a-8472-edaf3231c0b7" containerID="2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c" exitCode=0 Oct 02 07:29:44 crc kubenswrapper[4960]: I1002 07:29:44.814744 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgj42" event={"ID":"c7531862-540a-495a-8472-edaf3231c0b7","Type":"ContainerDied","Data":"2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c"} Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.119089 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.287084 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwblt\" (UniqueName: \"kubernetes.io/projected/b728cb6d-14fc-437a-abe1-836ee912d913-kube-api-access-gwblt\") pod \"b728cb6d-14fc-437a-abe1-836ee912d913\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.287199 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-util\") pod \"b728cb6d-14fc-437a-abe1-836ee912d913\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.287342 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-bundle\") pod \"b728cb6d-14fc-437a-abe1-836ee912d913\" (UID: \"b728cb6d-14fc-437a-abe1-836ee912d913\") " Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.289380 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-bundle" (OuterVolumeSpecName: "bundle") pod "b728cb6d-14fc-437a-abe1-836ee912d913" (UID: "b728cb6d-14fc-437a-abe1-836ee912d913"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.300362 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b728cb6d-14fc-437a-abe1-836ee912d913-kube-api-access-gwblt" (OuterVolumeSpecName: "kube-api-access-gwblt") pod "b728cb6d-14fc-437a-abe1-836ee912d913" (UID: "b728cb6d-14fc-437a-abe1-836ee912d913"). InnerVolumeSpecName "kube-api-access-gwblt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.389845 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwblt\" (UniqueName: \"kubernetes.io/projected/b728cb6d-14fc-437a-abe1-836ee912d913-kube-api-access-gwblt\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.389889 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.434425 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-util" (OuterVolumeSpecName: "util") pod "b728cb6d-14fc-437a-abe1-836ee912d913" (UID: "b728cb6d-14fc-437a-abe1-836ee912d913"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.491564 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b728cb6d-14fc-437a-abe1-836ee912d913-util\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.824172 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" event={"ID":"b728cb6d-14fc-437a-abe1-836ee912d913","Type":"ContainerDied","Data":"3cfb27391257a6ea851c43277ad9b43561a066051e6734c3b8dba3aa7df1929b"} Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.824225 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cfb27391257a6ea851c43277ad9b43561a066051e6734c3b8dba3aa7df1929b" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.824197 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s" Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.827298 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgj42" event={"ID":"c7531862-540a-495a-8472-edaf3231c0b7","Type":"ContainerStarted","Data":"3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57"} Oct 02 07:29:45 crc kubenswrapper[4960]: I1002 07:29:45.855379 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fgj42" podStartSLOduration=2.308980721 podStartE2EDuration="4.855359206s" podCreationTimestamp="2025-10-02 07:29:41 +0000 UTC" firstStartedPulling="2025-10-02 07:29:42.793699271 +0000 UTC m=+803.825645558" lastFinishedPulling="2025-10-02 07:29:45.340077756 +0000 UTC m=+806.372024043" observedRunningTime="2025-10-02 07:29:45.850936208 +0000 UTC m=+806.882882505" watchObservedRunningTime="2025-10-02 07:29:45.855359206 +0000 UTC m=+806.887305493" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.316680 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr"] Oct 02 07:29:48 crc kubenswrapper[4960]: E1002 07:29:48.317463 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b728cb6d-14fc-437a-abe1-836ee912d913" containerName="extract" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.317479 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b728cb6d-14fc-437a-abe1-836ee912d913" containerName="extract" Oct 02 07:29:48 crc kubenswrapper[4960]: E1002 07:29:48.317492 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b728cb6d-14fc-437a-abe1-836ee912d913" containerName="util" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.317500 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b728cb6d-14fc-437a-abe1-836ee912d913" containerName="util" Oct 02 07:29:48 crc kubenswrapper[4960]: E1002 07:29:48.317515 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b728cb6d-14fc-437a-abe1-836ee912d913" containerName="pull" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.317524 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b728cb6d-14fc-437a-abe1-836ee912d913" containerName="pull" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.317695 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b728cb6d-14fc-437a-abe1-836ee912d913" containerName="extract" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.318572 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.321414 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-zbpjz" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.346011 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr"] Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.435183 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrfss\" (UniqueName: \"kubernetes.io/projected/a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54-kube-api-access-wrfss\") pod \"openstack-operator-controller-operator-859455d779-rrmmr\" (UID: \"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54\") " pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.536726 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrfss\" (UniqueName: \"kubernetes.io/projected/a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54-kube-api-access-wrfss\") pod \"openstack-operator-controller-operator-859455d779-rrmmr\" (UID: \"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54\") " pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.560582 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrfss\" (UniqueName: \"kubernetes.io/projected/a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54-kube-api-access-wrfss\") pod \"openstack-operator-controller-operator-859455d779-rrmmr\" (UID: \"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54\") " pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 07:29:48 crc kubenswrapper[4960]: I1002 07:29:48.637639 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 07:29:49 crc kubenswrapper[4960]: I1002 07:29:49.136572 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr"] Oct 02 07:29:49 crc kubenswrapper[4960]: W1002 07:29:49.148235 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8fcd1e0_0c9e_453f_8e24_cfa92f46fe54.slice/crio-1b6b673893251c9c3e6ea00ef5fba1f4363862085dd9ec71ca6c0f7968a61bae WatchSource:0}: Error finding container 1b6b673893251c9c3e6ea00ef5fba1f4363862085dd9ec71ca6c0f7968a61bae: Status 404 returned error can't find the container with id 1b6b673893251c9c3e6ea00ef5fba1f4363862085dd9ec71ca6c0f7968a61bae Oct 02 07:29:49 crc kubenswrapper[4960]: I1002 07:29:49.865539 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" event={"ID":"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54","Type":"ContainerStarted","Data":"1b6b673893251c9c3e6ea00ef5fba1f4363862085dd9ec71ca6c0f7968a61bae"} Oct 02 07:29:52 crc kubenswrapper[4960]: I1002 07:29:52.138684 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:52 crc kubenswrapper[4960]: I1002 07:29:52.139488 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:52 crc kubenswrapper[4960]: I1002 07:29:52.189573 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:52 crc kubenswrapper[4960]: I1002 07:29:52.925789 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:53 crc kubenswrapper[4960]: I1002 07:29:53.896825 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" event={"ID":"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54","Type":"ContainerStarted","Data":"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf"} Oct 02 07:29:54 crc kubenswrapper[4960]: I1002 07:29:54.525270 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fgj42"] Oct 02 07:29:54 crc kubenswrapper[4960]: I1002 07:29:54.903726 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fgj42" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="registry-server" containerID="cri-o://3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57" gracePeriod=2 Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.667887 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.791918 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-utilities\") pod \"c7531862-540a-495a-8472-edaf3231c0b7\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.792025 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9ts8\" (UniqueName: \"kubernetes.io/projected/c7531862-540a-495a-8472-edaf3231c0b7-kube-api-access-n9ts8\") pod \"c7531862-540a-495a-8472-edaf3231c0b7\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.792163 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-catalog-content\") pod \"c7531862-540a-495a-8472-edaf3231c0b7\" (UID: \"c7531862-540a-495a-8472-edaf3231c0b7\") " Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.792962 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-utilities" (OuterVolumeSpecName: "utilities") pod "c7531862-540a-495a-8472-edaf3231c0b7" (UID: "c7531862-540a-495a-8472-edaf3231c0b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.801188 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7531862-540a-495a-8472-edaf3231c0b7-kube-api-access-n9ts8" (OuterVolumeSpecName: "kube-api-access-n9ts8") pod "c7531862-540a-495a-8472-edaf3231c0b7" (UID: "c7531862-540a-495a-8472-edaf3231c0b7"). InnerVolumeSpecName "kube-api-access-n9ts8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.894398 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.894448 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9ts8\" (UniqueName: \"kubernetes.io/projected/c7531862-540a-495a-8472-edaf3231c0b7-kube-api-access-n9ts8\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.898184 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7531862-540a-495a-8472-edaf3231c0b7" (UID: "c7531862-540a-495a-8472-edaf3231c0b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.914485 4960 generic.go:334] "Generic (PLEG): container finished" podID="c7531862-540a-495a-8472-edaf3231c0b7" containerID="3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57" exitCode=0 Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.914591 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgj42" event={"ID":"c7531862-540a-495a-8472-edaf3231c0b7","Type":"ContainerDied","Data":"3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57"} Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.914717 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgj42" event={"ID":"c7531862-540a-495a-8472-edaf3231c0b7","Type":"ContainerDied","Data":"ceb38bcc9b8b6a70cce0c34c1d9bbb62ab2f277a7c714d52c3c8aab2b3bfd072"} Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.914759 4960 scope.go:117] "RemoveContainer" containerID="3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.915227 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgj42" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.941535 4960 scope.go:117] "RemoveContainer" containerID="2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.956361 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fgj42"] Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.966395 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fgj42"] Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.969366 4960 scope.go:117] "RemoveContainer" containerID="d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e" Oct 02 07:29:55 crc kubenswrapper[4960]: I1002 07:29:55.995658 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7531862-540a-495a-8472-edaf3231c0b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:56 crc kubenswrapper[4960]: I1002 07:29:56.002143 4960 scope.go:117] "RemoveContainer" containerID="3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57" Oct 02 07:29:56 crc kubenswrapper[4960]: E1002 07:29:56.003283 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57\": container with ID starting with 3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57 not found: ID does not exist" containerID="3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57" Oct 02 07:29:56 crc kubenswrapper[4960]: I1002 07:29:56.003345 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57"} err="failed to get container status \"3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57\": rpc error: code = NotFound desc = could not find container \"3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57\": container with ID starting with 3dadaa4210000aa463cedfb776485dfffe9bb7f36c71abb36550e49809af2f57 not found: ID does not exist" Oct 02 07:29:56 crc kubenswrapper[4960]: I1002 07:29:56.003388 4960 scope.go:117] "RemoveContainer" containerID="2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c" Oct 02 07:29:56 crc kubenswrapper[4960]: E1002 07:29:56.003998 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c\": container with ID starting with 2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c not found: ID does not exist" containerID="2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c" Oct 02 07:29:56 crc kubenswrapper[4960]: I1002 07:29:56.004054 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c"} err="failed to get container status \"2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c\": rpc error: code = NotFound desc = could not find container \"2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c\": container with ID starting with 2c516fbf5c29fcb3ef28f3aab6aedb0bb3e5be85c870fbfc4a0a645b0350a14c not found: ID does not exist" Oct 02 07:29:56 crc kubenswrapper[4960]: I1002 07:29:56.004092 4960 scope.go:117] "RemoveContainer" containerID="d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e" Oct 02 07:29:56 crc kubenswrapper[4960]: E1002 07:29:56.011819 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e\": container with ID starting with d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e not found: ID does not exist" containerID="d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e" Oct 02 07:29:56 crc kubenswrapper[4960]: I1002 07:29:56.011878 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e"} err="failed to get container status \"d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e\": rpc error: code = NotFound desc = could not find container \"d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e\": container with ID starting with d4c3b9c38fc3433fa8bc762644f933133169b6185a8eb9adf6a25e9c8137903e not found: ID does not exist" Oct 02 07:29:56 crc kubenswrapper[4960]: I1002 07:29:56.341470 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7531862-540a-495a-8472-edaf3231c0b7" path="/var/lib/kubelet/pods/c7531862-540a-495a-8472-edaf3231c0b7/volumes" Oct 02 07:29:58 crc kubenswrapper[4960]: I1002 07:29:58.945922 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" event={"ID":"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54","Type":"ContainerStarted","Data":"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04"} Oct 02 07:29:58 crc kubenswrapper[4960]: I1002 07:29:58.946648 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 07:29:58 crc kubenswrapper[4960]: I1002 07:29:58.952971 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 07:29:58 crc kubenswrapper[4960]: I1002 07:29:58.980020 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" podStartSLOduration=2.103935005 podStartE2EDuration="10.979961746s" podCreationTimestamp="2025-10-02 07:29:48 +0000 UTC" firstStartedPulling="2025-10-02 07:29:49.153707798 +0000 UTC m=+810.185654085" lastFinishedPulling="2025-10-02 07:29:58.029734499 +0000 UTC m=+819.061680826" observedRunningTime="2025-10-02 07:29:58.97486383 +0000 UTC m=+820.006810197" watchObservedRunningTime="2025-10-02 07:29:58.979961746 +0000 UTC m=+820.011908063" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.123862 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k"] Oct 02 07:30:00 crc kubenswrapper[4960]: E1002 07:30:00.124180 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="extract-utilities" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.124197 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="extract-utilities" Oct 02 07:30:00 crc kubenswrapper[4960]: E1002 07:30:00.124216 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="registry-server" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.124223 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="registry-server" Oct 02 07:30:00 crc kubenswrapper[4960]: E1002 07:30:00.124241 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="extract-content" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.124249 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="extract-content" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.124357 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7531862-540a-495a-8472-edaf3231c0b7" containerName="registry-server" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.124843 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.127289 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.127581 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.135761 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k"] Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.158803 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-config-volume\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.158892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-secret-volume\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.159049 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr6dw\" (UniqueName: \"kubernetes.io/projected/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-kube-api-access-pr6dw\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.259741 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-config-volume\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.259869 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-secret-volume\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.259915 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr6dw\" (UniqueName: \"kubernetes.io/projected/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-kube-api-access-pr6dw\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.260745 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-config-volume\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.272315 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-secret-volume\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.278696 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr6dw\" (UniqueName: \"kubernetes.io/projected/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-kube-api-access-pr6dw\") pod \"collect-profiles-29323170-dzb9k\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.442394 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.889392 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k"] Oct 02 07:30:00 crc kubenswrapper[4960]: W1002 07:30:00.901827 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b65ad47_92d7_42f5_97d7_dc83810d1bc6.slice/crio-e39b8acc4f8175ef1d7968a35e6e325c8a57756ceddaeb4965079618ccacc7b7 WatchSource:0}: Error finding container e39b8acc4f8175ef1d7968a35e6e325c8a57756ceddaeb4965079618ccacc7b7: Status 404 returned error can't find the container with id e39b8acc4f8175ef1d7968a35e6e325c8a57756ceddaeb4965079618ccacc7b7 Oct 02 07:30:00 crc kubenswrapper[4960]: I1002 07:30:00.960137 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" event={"ID":"2b65ad47-92d7-42f5-97d7-dc83810d1bc6","Type":"ContainerStarted","Data":"e39b8acc4f8175ef1d7968a35e6e325c8a57756ceddaeb4965079618ccacc7b7"} Oct 02 07:30:01 crc kubenswrapper[4960]: I1002 07:30:01.968150 4960 generic.go:334] "Generic (PLEG): container finished" podID="2b65ad47-92d7-42f5-97d7-dc83810d1bc6" containerID="c89050772bc026f2534ff2a2ea5f3b2b9bfe53273cb1474a5081a7fd5d10adbb" exitCode=0 Oct 02 07:30:01 crc kubenswrapper[4960]: I1002 07:30:01.968386 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" event={"ID":"2b65ad47-92d7-42f5-97d7-dc83810d1bc6","Type":"ContainerDied","Data":"c89050772bc026f2534ff2a2ea5f3b2b9bfe53273cb1474a5081a7fd5d10adbb"} Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.258790 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.313747 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr6dw\" (UniqueName: \"kubernetes.io/projected/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-kube-api-access-pr6dw\") pod \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.313891 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-secret-volume\") pod \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.313929 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-config-volume\") pod \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\" (UID: \"2b65ad47-92d7-42f5-97d7-dc83810d1bc6\") " Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.314888 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-config-volume" (OuterVolumeSpecName: "config-volume") pod "2b65ad47-92d7-42f5-97d7-dc83810d1bc6" (UID: "2b65ad47-92d7-42f5-97d7-dc83810d1bc6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.319798 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2b65ad47-92d7-42f5-97d7-dc83810d1bc6" (UID: "2b65ad47-92d7-42f5-97d7-dc83810d1bc6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.319988 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-kube-api-access-pr6dw" (OuterVolumeSpecName: "kube-api-access-pr6dw") pod "2b65ad47-92d7-42f5-97d7-dc83810d1bc6" (UID: "2b65ad47-92d7-42f5-97d7-dc83810d1bc6"). InnerVolumeSpecName "kube-api-access-pr6dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.415737 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.416297 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.416322 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr6dw\" (UniqueName: \"kubernetes.io/projected/2b65ad47-92d7-42f5-97d7-dc83810d1bc6-kube-api-access-pr6dw\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.982721 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" event={"ID":"2b65ad47-92d7-42f5-97d7-dc83810d1bc6","Type":"ContainerDied","Data":"e39b8acc4f8175ef1d7968a35e6e325c8a57756ceddaeb4965079618ccacc7b7"} Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.982765 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e39b8acc4f8175ef1d7968a35e6e325c8a57756ceddaeb4965079618ccacc7b7" Oct 02 07:30:03 crc kubenswrapper[4960]: I1002 07:30:03.982800 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.599197 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kx747"] Oct 02 07:30:05 crc kubenswrapper[4960]: E1002 07:30:05.599479 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b65ad47-92d7-42f5-97d7-dc83810d1bc6" containerName="collect-profiles" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.599490 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b65ad47-92d7-42f5-97d7-dc83810d1bc6" containerName="collect-profiles" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.599611 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b65ad47-92d7-42f5-97d7-dc83810d1bc6" containerName="collect-profiles" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.600800 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.615454 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx747"] Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.745524 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-catalog-content\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.745612 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-utilities\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.745652 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rtsp\" (UniqueName: \"kubernetes.io/projected/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-kube-api-access-6rtsp\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.847607 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-utilities\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.847670 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rtsp\" (UniqueName: \"kubernetes.io/projected/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-kube-api-access-6rtsp\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.847745 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-catalog-content\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.848451 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-catalog-content\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.848484 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-utilities\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.873385 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rtsp\" (UniqueName: \"kubernetes.io/projected/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-kube-api-access-6rtsp\") pod \"redhat-marketplace-kx747\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:05 crc kubenswrapper[4960]: I1002 07:30:05.917768 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:06 crc kubenswrapper[4960]: I1002 07:30:06.201593 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx747"] Oct 02 07:30:06 crc kubenswrapper[4960]: W1002 07:30:06.207241 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1fba335_8b2a_4f7c_a78f_6e5b3a2cfee9.slice/crio-17b82905f44b5fa83c1b4f9c977c00f31ebf8ffb8a7307ffdb18c077b7308047 WatchSource:0}: Error finding container 17b82905f44b5fa83c1b4f9c977c00f31ebf8ffb8a7307ffdb18c077b7308047: Status 404 returned error can't find the container with id 17b82905f44b5fa83c1b4f9c977c00f31ebf8ffb8a7307ffdb18c077b7308047 Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.002461 4960 generic.go:334] "Generic (PLEG): container finished" podID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerID="a10b75b9fb1be64790ab2dbfb0b0d7f06b456e1eb7bf90b420146b5292353845" exitCode=0 Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.002517 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx747" event={"ID":"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9","Type":"ContainerDied","Data":"a10b75b9fb1be64790ab2dbfb0b0d7f06b456e1eb7bf90b420146b5292353845"} Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.002550 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx747" event={"ID":"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9","Type":"ContainerStarted","Data":"17b82905f44b5fa83c1b4f9c977c00f31ebf8ffb8a7307ffdb18c077b7308047"} Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.951667 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j9pkx"] Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.953861 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.964404 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9pkx"] Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.981274 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e06b264d-7f1e-48b9-a067-1d428f2633f4-catalog-content\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.981354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gspng\" (UniqueName: \"kubernetes.io/projected/e06b264d-7f1e-48b9-a067-1d428f2633f4-kube-api-access-gspng\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:07 crc kubenswrapper[4960]: I1002 07:30:07.981391 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e06b264d-7f1e-48b9-a067-1d428f2633f4-utilities\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.011130 4960 generic.go:334] "Generic (PLEG): container finished" podID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerID="f4d03ef8741b23acb29c81dc03b5462f69d601ef7e8eb49aaf931265c274411d" exitCode=0 Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.011192 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx747" event={"ID":"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9","Type":"ContainerDied","Data":"f4d03ef8741b23acb29c81dc03b5462f69d601ef7e8eb49aaf931265c274411d"} Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.082914 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e06b264d-7f1e-48b9-a067-1d428f2633f4-catalog-content\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.082968 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gspng\" (UniqueName: \"kubernetes.io/projected/e06b264d-7f1e-48b9-a067-1d428f2633f4-kube-api-access-gspng\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.083051 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e06b264d-7f1e-48b9-a067-1d428f2633f4-utilities\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.083626 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e06b264d-7f1e-48b9-a067-1d428f2633f4-catalog-content\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.083670 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e06b264d-7f1e-48b9-a067-1d428f2633f4-utilities\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.106670 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gspng\" (UniqueName: \"kubernetes.io/projected/e06b264d-7f1e-48b9-a067-1d428f2633f4-kube-api-access-gspng\") pod \"community-operators-j9pkx\" (UID: \"e06b264d-7f1e-48b9-a067-1d428f2633f4\") " pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.271054 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:08 crc kubenswrapper[4960]: I1002 07:30:08.807232 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9pkx"] Oct 02 07:30:08 crc kubenswrapper[4960]: W1002 07:30:08.827802 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode06b264d_7f1e_48b9_a067_1d428f2633f4.slice/crio-8e0d6f33df271538e239f9c1ea600f4ae533225a6d8bae0fab35a4829368f748 WatchSource:0}: Error finding container 8e0d6f33df271538e239f9c1ea600f4ae533225a6d8bae0fab35a4829368f748: Status 404 returned error can't find the container with id 8e0d6f33df271538e239f9c1ea600f4ae533225a6d8bae0fab35a4829368f748 Oct 02 07:30:09 crc kubenswrapper[4960]: I1002 07:30:09.017948 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx747" event={"ID":"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9","Type":"ContainerStarted","Data":"05fa352a898b8ce0074aa520b060a8046651106341e98697dc9b7615cd9de27e"} Oct 02 07:30:09 crc kubenswrapper[4960]: I1002 07:30:09.019669 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9pkx" event={"ID":"e06b264d-7f1e-48b9-a067-1d428f2633f4","Type":"ContainerStarted","Data":"8e0d6f33df271538e239f9c1ea600f4ae533225a6d8bae0fab35a4829368f748"} Oct 02 07:30:09 crc kubenswrapper[4960]: I1002 07:30:09.037323 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kx747" podStartSLOduration=2.54656909 podStartE2EDuration="4.037307954s" podCreationTimestamp="2025-10-02 07:30:05 +0000 UTC" firstStartedPulling="2025-10-02 07:30:07.003773573 +0000 UTC m=+828.035719860" lastFinishedPulling="2025-10-02 07:30:08.494512447 +0000 UTC m=+829.526458724" observedRunningTime="2025-10-02 07:30:09.036757918 +0000 UTC m=+830.068704205" watchObservedRunningTime="2025-10-02 07:30:09.037307954 +0000 UTC m=+830.069254241" Oct 02 07:30:10 crc kubenswrapper[4960]: I1002 07:30:10.028902 4960 generic.go:334] "Generic (PLEG): container finished" podID="e06b264d-7f1e-48b9-a067-1d428f2633f4" containerID="06400afce5793ca62044cd7a8b3a466e59925c88beb73070ed2e081e3e56a8c3" exitCode=0 Oct 02 07:30:10 crc kubenswrapper[4960]: I1002 07:30:10.028967 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9pkx" event={"ID":"e06b264d-7f1e-48b9-a067-1d428f2633f4","Type":"ContainerDied","Data":"06400afce5793ca62044cd7a8b3a466e59925c88beb73070ed2e081e3e56a8c3"} Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.408876 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.410781 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.413174 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-lv298" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.434537 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.468197 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.469824 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.474430 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-xrfd2" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.475169 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.476739 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.478490 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qldln" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.490668 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.497958 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.512037 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.513403 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.514849 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9sgp\" (UniqueName: \"kubernetes.io/projected/b6b8a229-03cf-4cb0-ab8a-a3133dadfc21-kube-api-access-k9sgp\") pod \"barbican-operator-controller-manager-f7f98cb69-mfg89\" (UID: \"b6b8a229-03cf-4cb0-ab8a-a3133dadfc21\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.517365 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-h6b6s" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.521119 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.526942 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.528189 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.532084 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-jg2mb" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.542772 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.543911 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.551392 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-thddj" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.551920 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.562996 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.564409 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.566427 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.566757 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-7kklc" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.569590 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.583937 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.585621 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.590356 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-6qr6q" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.624766 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9sgp\" (UniqueName: \"kubernetes.io/projected/b6b8a229-03cf-4cb0-ab8a-a3133dadfc21-kube-api-access-k9sgp\") pod \"barbican-operator-controller-manager-f7f98cb69-mfg89\" (UID: \"b6b8a229-03cf-4cb0-ab8a-a3133dadfc21\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.624851 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67m26\" (UniqueName: \"kubernetes.io/projected/b9dd3b02-9840-4eef-b378-b7888406a57c-kube-api-access-67m26\") pod \"glance-operator-controller-manager-8bc4775b5-5mk92\" (UID: \"b9dd3b02-9840-4eef-b378-b7888406a57c\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.626067 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pjgr\" (UniqueName: \"kubernetes.io/projected/539fe2a2-8989-4fb6-a24e-dee67304e96b-kube-api-access-2pjgr\") pod \"cinder-operator-controller-manager-859cd486d-dfpnm\" (UID: \"539fe2a2-8989-4fb6-a24e-dee67304e96b\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.626149 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6885\" (UniqueName: \"kubernetes.io/projected/6c71ae76-8a1e-4570-8950-8cf4962169e4-kube-api-access-p6885\") pod \"designate-operator-controller-manager-77fb7bcf5b-7qjq7\" (UID: \"6c71ae76-8a1e-4570-8950-8cf4962169e4\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.656842 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.682206 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9sgp\" (UniqueName: \"kubernetes.io/projected/b6b8a229-03cf-4cb0-ab8a-a3133dadfc21-kube-api-access-k9sgp\") pod \"barbican-operator-controller-manager-f7f98cb69-mfg89\" (UID: \"b6b8a229-03cf-4cb0-ab8a-a3133dadfc21\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.719892 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.721057 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.724489 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.728861 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67m26\" (UniqueName: \"kubernetes.io/projected/b9dd3b02-9840-4eef-b378-b7888406a57c-kube-api-access-67m26\") pod \"glance-operator-controller-manager-8bc4775b5-5mk92\" (UID: \"b9dd3b02-9840-4eef-b378-b7888406a57c\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.728919 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pjgr\" (UniqueName: \"kubernetes.io/projected/539fe2a2-8989-4fb6-a24e-dee67304e96b-kube-api-access-2pjgr\") pod \"cinder-operator-controller-manager-859cd486d-dfpnm\" (UID: \"539fe2a2-8989-4fb6-a24e-dee67304e96b\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.728946 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6885\" (UniqueName: \"kubernetes.io/projected/6c71ae76-8a1e-4570-8950-8cf4962169e4-kube-api-access-p6885\") pod \"designate-operator-controller-manager-77fb7bcf5b-7qjq7\" (UID: \"6c71ae76-8a1e-4570-8950-8cf4962169e4\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.729009 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nflq2\" (UniqueName: \"kubernetes.io/projected/cf515440-a75b-4f32-8c08-2f108da16b56-kube-api-access-nflq2\") pod \"infra-operator-controller-manager-5c8fdc4d5c-hbzp7\" (UID: \"cf515440-a75b-4f32-8c08-2f108da16b56\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.729034 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf515440-a75b-4f32-8c08-2f108da16b56-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-hbzp7\" (UID: \"cf515440-a75b-4f32-8c08-2f108da16b56\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.729054 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvjm6\" (UniqueName: \"kubernetes.io/projected/72b0c3f8-2a66-42da-aa83-298bda8f332d-kube-api-access-tvjm6\") pod \"horizon-operator-controller-manager-679b4759bb-cd9cz\" (UID: \"72b0c3f8-2a66-42da-aa83-298bda8f332d\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.729082 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n2t7\" (UniqueName: \"kubernetes.io/projected/32ba83c2-5436-4b71-9a95-e552613c6bff-kube-api-access-7n2t7\") pod \"heat-operator-controller-manager-5b4fc86755-g9bwh\" (UID: \"32ba83c2-5436-4b71-9a95-e552613c6bff\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.729106 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz5sj\" (UniqueName: \"kubernetes.io/projected/4bcfb151-7975-4567-9a1f-0cb7dd128696-kube-api-access-qz5sj\") pod \"ironic-operator-controller-manager-5f45cd594f-wm4wc\" (UID: \"4bcfb151-7975-4567-9a1f-0cb7dd128696\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.732742 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-l22h6" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.750358 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.757349 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.762054 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.770213 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.788450 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-dlbjr" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.789009 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.789406 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pjgr\" (UniqueName: \"kubernetes.io/projected/539fe2a2-8989-4fb6-a24e-dee67304e96b-kube-api-access-2pjgr\") pod \"cinder-operator-controller-manager-859cd486d-dfpnm\" (UID: \"539fe2a2-8989-4fb6-a24e-dee67304e96b\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.802032 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67m26\" (UniqueName: \"kubernetes.io/projected/b9dd3b02-9840-4eef-b378-b7888406a57c-kube-api-access-67m26\") pod \"glance-operator-controller-manager-8bc4775b5-5mk92\" (UID: \"b9dd3b02-9840-4eef-b378-b7888406a57c\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.802574 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6885\" (UniqueName: \"kubernetes.io/projected/6c71ae76-8a1e-4570-8950-8cf4962169e4-kube-api-access-p6885\") pod \"designate-operator-controller-manager-77fb7bcf5b-7qjq7\" (UID: \"6c71ae76-8a1e-4570-8950-8cf4962169e4\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.810012 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.812052 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.821481 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2q8zt" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.836262 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nflq2\" (UniqueName: \"kubernetes.io/projected/cf515440-a75b-4f32-8c08-2f108da16b56-kube-api-access-nflq2\") pod \"infra-operator-controller-manager-5c8fdc4d5c-hbzp7\" (UID: \"cf515440-a75b-4f32-8c08-2f108da16b56\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.836316 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf515440-a75b-4f32-8c08-2f108da16b56-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-hbzp7\" (UID: \"cf515440-a75b-4f32-8c08-2f108da16b56\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.836339 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvjm6\" (UniqueName: \"kubernetes.io/projected/72b0c3f8-2a66-42da-aa83-298bda8f332d-kube-api-access-tvjm6\") pod \"horizon-operator-controller-manager-679b4759bb-cd9cz\" (UID: \"72b0c3f8-2a66-42da-aa83-298bda8f332d\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.836370 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n2t7\" (UniqueName: \"kubernetes.io/projected/32ba83c2-5436-4b71-9a95-e552613c6bff-kube-api-access-7n2t7\") pod \"heat-operator-controller-manager-5b4fc86755-g9bwh\" (UID: \"32ba83c2-5436-4b71-9a95-e552613c6bff\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.836395 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz5sj\" (UniqueName: \"kubernetes.io/projected/4bcfb151-7975-4567-9a1f-0cb7dd128696-kube-api-access-qz5sj\") pod \"ironic-operator-controller-manager-5f45cd594f-wm4wc\" (UID: \"4bcfb151-7975-4567-9a1f-0cb7dd128696\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.836440 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcsqx\" (UniqueName: \"kubernetes.io/projected/8c352774-9d31-4d0a-a1ee-0a96bfbf6239-kube-api-access-pcsqx\") pod \"keystone-operator-controller-manager-59d7dc95cf-54crk\" (UID: \"8c352774-9d31-4d0a-a1ee-0a96bfbf6239\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.836791 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" Oct 02 07:30:15 crc kubenswrapper[4960]: E1002 07:30:15.837749 4960 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 02 07:30:15 crc kubenswrapper[4960]: E1002 07:30:15.837815 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf515440-a75b-4f32-8c08-2f108da16b56-cert podName:cf515440-a75b-4f32-8c08-2f108da16b56 nodeName:}" failed. No retries permitted until 2025-10-02 07:30:16.337793588 +0000 UTC m=+837.369739875 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf515440-a75b-4f32-8c08-2f108da16b56-cert") pod "infra-operator-controller-manager-5c8fdc4d5c-hbzp7" (UID: "cf515440-a75b-4f32-8c08-2f108da16b56") : secret "infra-operator-webhook-server-cert" not found Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.848000 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.863895 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.901764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nflq2\" (UniqueName: \"kubernetes.io/projected/cf515440-a75b-4f32-8c08-2f108da16b56-kube-api-access-nflq2\") pod \"infra-operator-controller-manager-5c8fdc4d5c-hbzp7\" (UID: \"cf515440-a75b-4f32-8c08-2f108da16b56\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.902543 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz5sj\" (UniqueName: \"kubernetes.io/projected/4bcfb151-7975-4567-9a1f-0cb7dd128696-kube-api-access-qz5sj\") pod \"ironic-operator-controller-manager-5f45cd594f-wm4wc\" (UID: \"4bcfb151-7975-4567-9a1f-0cb7dd128696\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.909841 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n2t7\" (UniqueName: \"kubernetes.io/projected/32ba83c2-5436-4b71-9a95-e552613c6bff-kube-api-access-7n2t7\") pod \"heat-operator-controller-manager-5b4fc86755-g9bwh\" (UID: \"32ba83c2-5436-4b71-9a95-e552613c6bff\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.918386 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.919279 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.919781 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvjm6\" (UniqueName: \"kubernetes.io/projected/72b0c3f8-2a66-42da-aa83-298bda8f332d-kube-api-access-tvjm6\") pod \"horizon-operator-controller-manager-679b4759bb-cd9cz\" (UID: \"72b0c3f8-2a66-42da-aa83-298bda8f332d\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.932190 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.934742 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.948678 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll884\" (UniqueName: \"kubernetes.io/projected/2af8b0c5-6356-4c12-894e-1bb39e4bf981-kube-api-access-ll884\") pod \"manila-operator-controller-manager-b7cf8cb5f-hkswg\" (UID: \"2af8b0c5-6356-4c12-894e-1bb39e4bf981\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.948743 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcsqx\" (UniqueName: \"kubernetes.io/projected/8c352774-9d31-4d0a-a1ee-0a96bfbf6239-kube-api-access-pcsqx\") pod \"keystone-operator-controller-manager-59d7dc95cf-54crk\" (UID: \"8c352774-9d31-4d0a-a1ee-0a96bfbf6239\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.948787 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb6xl\" (UniqueName: \"kubernetes.io/projected/5d7077dd-5b52-49af-9dd7-c81f361139ec-kube-api-access-zb6xl\") pod \"mariadb-operator-controller-manager-67bf5bb885-kgzd2\" (UID: \"5d7077dd-5b52-49af-9dd7-c81f361139ec\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.965098 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs"] Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.966469 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.972927 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-df6nl" Oct 02 07:30:15 crc kubenswrapper[4960]: I1002 07:30:15.974302 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcsqx\" (UniqueName: \"kubernetes.io/projected/8c352774-9d31-4d0a-a1ee-0a96bfbf6239-kube-api-access-pcsqx\") pod \"keystone-operator-controller-manager-59d7dc95cf-54crk\" (UID: \"8c352774-9d31-4d0a-a1ee-0a96bfbf6239\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.015589 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.017934 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.030490 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-tm24t" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.034876 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.049826 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb6xl\" (UniqueName: \"kubernetes.io/projected/5d7077dd-5b52-49af-9dd7-c81f361139ec-kube-api-access-zb6xl\") pod \"mariadb-operator-controller-manager-67bf5bb885-kgzd2\" (UID: \"5d7077dd-5b52-49af-9dd7-c81f361139ec\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.049923 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll884\" (UniqueName: \"kubernetes.io/projected/2af8b0c5-6356-4c12-894e-1bb39e4bf981-kube-api-access-ll884\") pod \"manila-operator-controller-manager-b7cf8cb5f-hkswg\" (UID: \"2af8b0c5-6356-4c12-894e-1bb39e4bf981\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.061255 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.069029 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.070433 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.094210 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.096483 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll884\" (UniqueName: \"kubernetes.io/projected/2af8b0c5-6356-4c12-894e-1bb39e4bf981-kube-api-access-ll884\") pod \"manila-operator-controller-manager-b7cf8cb5f-hkswg\" (UID: \"2af8b0c5-6356-4c12-894e-1bb39e4bf981\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.098813 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-ztjwp" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.108099 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb6xl\" (UniqueName: \"kubernetes.io/projected/5d7077dd-5b52-49af-9dd7-c81f361139ec-kube-api-access-zb6xl\") pod \"mariadb-operator-controller-manager-67bf5bb885-kgzd2\" (UID: \"5d7077dd-5b52-49af-9dd7-c81f361139ec\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.113802 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.142050 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.153482 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7whv\" (UniqueName: \"kubernetes.io/projected/418a02f4-eb9e-4079-969a-c75f8307b078-kube-api-access-w7whv\") pod \"octavia-operator-controller-manager-75f8d67d86-xhd8b\" (UID: \"418a02f4-eb9e-4079-969a-c75f8307b078\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.153526 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdr76\" (UniqueName: \"kubernetes.io/projected/c0d6bd86-ab92-4f37-be9c-003aab10f910-kube-api-access-kdr76\") pod \"neutron-operator-controller-manager-54fbbfcd44-6lpjs\" (UID: \"c0d6bd86-ab92-4f37-be9c-003aab10f910\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.153645 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8xx6\" (UniqueName: \"kubernetes.io/projected/87d511d4-0cbf-49b2-a2e2-b2e2fc726f60-kube-api-access-s8xx6\") pod \"nova-operator-controller-manager-7fd5b6bbc6-4vtmz\" (UID: \"87d511d4-0cbf-49b2-a2e2-b2e2fc726f60\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.167834 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9pkx" event={"ID":"e06b264d-7f1e-48b9-a067-1d428f2633f4","Type":"ContainerStarted","Data":"bbc1fe0473e72aad84c3b5b02f3ec95045973f7bb5af1fb4a5270069d2b4f72d"} Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.176442 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.189605 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.209395 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.237391 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.239164 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.240781 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.241424 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.246435 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8ndzx" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.246616 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-xmpgc" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.247647 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.258182 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh8q6\" (UniqueName: \"kubernetes.io/projected/a786a744-833c-40f9-91a7-8b5f312a626c-kube-api-access-kh8q6\") pod \"ovn-operator-controller-manager-84c745747f-lk5kd\" (UID: \"a786a744-833c-40f9-91a7-8b5f312a626c\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.258281 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70d6affd-3067-440a-8b1e-69285828be97-cert\") pod \"openstack-baremetal-operator-controller-manager-787874f5b776jcd\" (UID: \"70d6affd-3067-440a-8b1e-69285828be97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.258309 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g8jw\" (UniqueName: \"kubernetes.io/projected/70d6affd-3067-440a-8b1e-69285828be97-kube-api-access-9g8jw\") pod \"openstack-baremetal-operator-controller-manager-787874f5b776jcd\" (UID: \"70d6affd-3067-440a-8b1e-69285828be97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.258346 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8xx6\" (UniqueName: \"kubernetes.io/projected/87d511d4-0cbf-49b2-a2e2-b2e2fc726f60-kube-api-access-s8xx6\") pod \"nova-operator-controller-manager-7fd5b6bbc6-4vtmz\" (UID: \"87d511d4-0cbf-49b2-a2e2-b2e2fc726f60\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.258373 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7whv\" (UniqueName: \"kubernetes.io/projected/418a02f4-eb9e-4079-969a-c75f8307b078-kube-api-access-w7whv\") pod \"octavia-operator-controller-manager-75f8d67d86-xhd8b\" (UID: \"418a02f4-eb9e-4079-969a-c75f8307b078\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.258394 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdr76\" (UniqueName: \"kubernetes.io/projected/c0d6bd86-ab92-4f37-be9c-003aab10f910-kube-api-access-kdr76\") pod \"neutron-operator-controller-manager-54fbbfcd44-6lpjs\" (UID: \"c0d6bd86-ab92-4f37-be9c-003aab10f910\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.263759 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.265333 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.267595 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-89qgc" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.278996 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.295221 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8xx6\" (UniqueName: \"kubernetes.io/projected/87d511d4-0cbf-49b2-a2e2-b2e2fc726f60-kube-api-access-s8xx6\") pod \"nova-operator-controller-manager-7fd5b6bbc6-4vtmz\" (UID: \"87d511d4-0cbf-49b2-a2e2-b2e2fc726f60\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.301698 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdr76\" (UniqueName: \"kubernetes.io/projected/c0d6bd86-ab92-4f37-be9c-003aab10f910-kube-api-access-kdr76\") pod \"neutron-operator-controller-manager-54fbbfcd44-6lpjs\" (UID: \"c0d6bd86-ab92-4f37-be9c-003aab10f910\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.301813 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.306903 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.315482 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-8vz8t" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.317309 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.327514 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.330831 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.334763 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-sf6zn" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.372936 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7whv\" (UniqueName: \"kubernetes.io/projected/418a02f4-eb9e-4079-969a-c75f8307b078-kube-api-access-w7whv\") pod \"octavia-operator-controller-manager-75f8d67d86-xhd8b\" (UID: \"418a02f4-eb9e-4079-969a-c75f8307b078\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.377648 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.379389 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.397849 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf515440-a75b-4f32-8c08-2f108da16b56-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-hbzp7\" (UID: \"cf515440-a75b-4f32-8c08-2f108da16b56\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.397940 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh8q6\" (UniqueName: \"kubernetes.io/projected/a786a744-833c-40f9-91a7-8b5f312a626c-kube-api-access-kh8q6\") pod \"ovn-operator-controller-manager-84c745747f-lk5kd\" (UID: \"a786a744-833c-40f9-91a7-8b5f312a626c\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.398065 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70d6affd-3067-440a-8b1e-69285828be97-cert\") pod \"openstack-baremetal-operator-controller-manager-787874f5b776jcd\" (UID: \"70d6affd-3067-440a-8b1e-69285828be97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.398092 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g8jw\" (UniqueName: \"kubernetes.io/projected/70d6affd-3067-440a-8b1e-69285828be97-kube-api-access-9g8jw\") pod \"openstack-baremetal-operator-controller-manager-787874f5b776jcd\" (UID: \"70d6affd-3067-440a-8b1e-69285828be97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: E1002 07:30:16.399722 4960 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 07:30:16 crc kubenswrapper[4960]: E1002 07:30:16.399788 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70d6affd-3067-440a-8b1e-69285828be97-cert podName:70d6affd-3067-440a-8b1e-69285828be97 nodeName:}" failed. No retries permitted until 2025-10-02 07:30:16.899769128 +0000 UTC m=+837.931715415 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70d6affd-3067-440a-8b1e-69285828be97-cert") pod "openstack-baremetal-operator-controller-manager-787874f5b776jcd" (UID: "70d6affd-3067-440a-8b1e-69285828be97") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.421264 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf515440-a75b-4f32-8c08-2f108da16b56-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-hbzp7\" (UID: \"cf515440-a75b-4f32-8c08-2f108da16b56\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.454481 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh8q6\" (UniqueName: \"kubernetes.io/projected/a786a744-833c-40f9-91a7-8b5f312a626c-kube-api-access-kh8q6\") pod \"ovn-operator-controller-manager-84c745747f-lk5kd\" (UID: \"a786a744-833c-40f9-91a7-8b5f312a626c\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.502836 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g8jw\" (UniqueName: \"kubernetes.io/projected/70d6affd-3067-440a-8b1e-69285828be97-kube-api-access-9g8jw\") pod \"openstack-baremetal-operator-controller-manager-787874f5b776jcd\" (UID: \"70d6affd-3067-440a-8b1e-69285828be97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.504531 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.505943 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzzbj\" (UniqueName: \"kubernetes.io/projected/e52d2a9e-b404-44cf-978a-a2fed39d6337-kube-api-access-qzzbj\") pod \"swift-operator-controller-manager-689b4f76c9-lzlkk\" (UID: \"e52d2a9e-b404-44cf-978a-a2fed39d6337\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.506015 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ththn\" (UniqueName: \"kubernetes.io/projected/de18430c-640c-46ac-9eb5-de6216332a80-kube-api-access-ththn\") pod \"placement-operator-controller-manager-598c4c8547-fkgnk\" (UID: \"de18430c-640c-46ac-9eb5-de6216332a80\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.506067 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4lck\" (UniqueName: \"kubernetes.io/projected/73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894-kube-api-access-n4lck\") pod \"telemetry-operator-controller-manager-cb66d6b59-cmrs5\" (UID: \"73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.530037 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.536985 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.537061 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.537075 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.537091 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.537107 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.538409 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.538505 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.539395 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.554542 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.556609 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.563356 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-g86bk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.564238 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5xb8g" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.583993 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.598955 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.609678 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4lck\" (UniqueName: \"kubernetes.io/projected/73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894-kube-api-access-n4lck\") pod \"telemetry-operator-controller-manager-cb66d6b59-cmrs5\" (UID: \"73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.609772 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzzbj\" (UniqueName: \"kubernetes.io/projected/e52d2a9e-b404-44cf-978a-a2fed39d6337-kube-api-access-qzzbj\") pod \"swift-operator-controller-manager-689b4f76c9-lzlkk\" (UID: \"e52d2a9e-b404-44cf-978a-a2fed39d6337\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.609811 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ththn\" (UniqueName: \"kubernetes.io/projected/de18430c-640c-46ac-9eb5-de6216332a80-kube-api-access-ththn\") pod \"placement-operator-controller-manager-598c4c8547-fkgnk\" (UID: \"de18430c-640c-46ac-9eb5-de6216332a80\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.636909 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ththn\" (UniqueName: \"kubernetes.io/projected/de18430c-640c-46ac-9eb5-de6216332a80-kube-api-access-ththn\") pod \"placement-operator-controller-manager-598c4c8547-fkgnk\" (UID: \"de18430c-640c-46ac-9eb5-de6216332a80\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.642155 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4lck\" (UniqueName: \"kubernetes.io/projected/73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894-kube-api-access-n4lck\") pod \"telemetry-operator-controller-manager-cb66d6b59-cmrs5\" (UID: \"73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.644742 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzzbj\" (UniqueName: \"kubernetes.io/projected/e52d2a9e-b404-44cf-978a-a2fed39d6337-kube-api-access-qzzbj\") pod \"swift-operator-controller-manager-689b4f76c9-lzlkk\" (UID: \"e52d2a9e-b404-44cf-978a-a2fed39d6337\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.653047 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.664031 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.698460 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.700023 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.713630 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-s2wl2" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.713860 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.714017 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.716981 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-strzj\" (UniqueName: \"kubernetes.io/projected/97605d31-7ce2-4f75-848a-1fcf39e54d3e-kube-api-access-strzj\") pod \"watcher-operator-controller-manager-68d7bc5569-69k7g\" (UID: \"97605d31-7ce2-4f75-848a-1fcf39e54d3e\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.717069 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l7sf\" (UniqueName: \"kubernetes.io/projected/5529d258-1683-402a-ba4d-fa355405201d-kube-api-access-4l7sf\") pod \"test-operator-controller-manager-cbdf6dc66-sqb6w\" (UID: \"5529d258-1683-402a-ba4d-fa355405201d\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.735368 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.736522 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.741447 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-hpz8w" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.755046 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.823394 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l7sf\" (UniqueName: \"kubernetes.io/projected/5529d258-1683-402a-ba4d-fa355405201d-kube-api-access-4l7sf\") pod \"test-operator-controller-manager-cbdf6dc66-sqb6w\" (UID: \"5529d258-1683-402a-ba4d-fa355405201d\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.823550 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f802e6c8-39a5-4acf-b85c-29f235283937-cert\") pod \"openstack-operator-controller-manager-67698bcd47-nz8r6\" (UID: \"f802e6c8-39a5-4acf-b85c-29f235283937\") " pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.823590 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7jlk\" (UniqueName: \"kubernetes.io/projected/f802e6c8-39a5-4acf-b85c-29f235283937-kube-api-access-v7jlk\") pod \"openstack-operator-controller-manager-67698bcd47-nz8r6\" (UID: \"f802e6c8-39a5-4acf-b85c-29f235283937\") " pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.823627 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-strzj\" (UniqueName: \"kubernetes.io/projected/97605d31-7ce2-4f75-848a-1fcf39e54d3e-kube-api-access-strzj\") pod \"watcher-operator-controller-manager-68d7bc5569-69k7g\" (UID: \"97605d31-7ce2-4f75-848a-1fcf39e54d3e\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.841879 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.846924 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-strzj\" (UniqueName: \"kubernetes.io/projected/97605d31-7ce2-4f75-848a-1fcf39e54d3e-kube-api-access-strzj\") pod \"watcher-operator-controller-manager-68d7bc5569-69k7g\" (UID: \"97605d31-7ce2-4f75-848a-1fcf39e54d3e\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.853408 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l7sf\" (UniqueName: \"kubernetes.io/projected/5529d258-1683-402a-ba4d-fa355405201d-kube-api-access-4l7sf\") pod \"test-operator-controller-manager-cbdf6dc66-sqb6w\" (UID: \"5529d258-1683-402a-ba4d-fa355405201d\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.881643 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.887480 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm"] Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.924796 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70d6affd-3067-440a-8b1e-69285828be97-cert\") pod \"openstack-baremetal-operator-controller-manager-787874f5b776jcd\" (UID: \"70d6affd-3067-440a-8b1e-69285828be97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.924870 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cjk4\" (UniqueName: \"kubernetes.io/projected/2fac2998-8d33-448f-877a-388b6570a141-kube-api-access-9cjk4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk\" (UID: \"2fac2998-8d33-448f-877a-388b6570a141\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.924915 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f802e6c8-39a5-4acf-b85c-29f235283937-cert\") pod \"openstack-operator-controller-manager-67698bcd47-nz8r6\" (UID: \"f802e6c8-39a5-4acf-b85c-29f235283937\") " pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.924951 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7jlk\" (UniqueName: \"kubernetes.io/projected/f802e6c8-39a5-4acf-b85c-29f235283937-kube-api-access-v7jlk\") pod \"openstack-operator-controller-manager-67698bcd47-nz8r6\" (UID: \"f802e6c8-39a5-4acf-b85c-29f235283937\") " pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:16 crc kubenswrapper[4960]: E1002 07:30:16.925248 4960 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 07:30:16 crc kubenswrapper[4960]: E1002 07:30:16.925370 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f802e6c8-39a5-4acf-b85c-29f235283937-cert podName:f802e6c8-39a5-4acf-b85c-29f235283937 nodeName:}" failed. No retries permitted until 2025-10-02 07:30:17.425345033 +0000 UTC m=+838.457291320 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f802e6c8-39a5-4acf-b85c-29f235283937-cert") pod "openstack-operator-controller-manager-67698bcd47-nz8r6" (UID: "f802e6c8-39a5-4acf-b85c-29f235283937") : secret "webhook-server-cert" not found Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.933223 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70d6affd-3067-440a-8b1e-69285828be97-cert\") pod \"openstack-baremetal-operator-controller-manager-787874f5b776jcd\" (UID: \"70d6affd-3067-440a-8b1e-69285828be97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.945205 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7jlk\" (UniqueName: \"kubernetes.io/projected/f802e6c8-39a5-4acf-b85c-29f235283937-kube-api-access-v7jlk\") pod \"openstack-operator-controller-manager-67698bcd47-nz8r6\" (UID: \"f802e6c8-39a5-4acf-b85c-29f235283937\") " pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.973615 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 07:30:16 crc kubenswrapper[4960]: I1002 07:30:16.980481 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.010423 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.026316 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cjk4\" (UniqueName: \"kubernetes.io/projected/2fac2998-8d33-448f-877a-388b6570a141-kube-api-access-9cjk4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk\" (UID: \"2fac2998-8d33-448f-877a-388b6570a141\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.028949 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.058993 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cjk4\" (UniqueName: \"kubernetes.io/projected/2fac2998-8d33-448f-877a-388b6570a141-kube-api-access-9cjk4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk\" (UID: \"2fac2998-8d33-448f-877a-388b6570a141\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.095855 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.112438 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.142999 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7"] Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.203630 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c71ae76_8a1e_4570_8950_8cf4962169e4.slice/crio-c175c5d68cca3fce2bd6848cac93db854bf7d126ba83861351b14e2aa46b056a WatchSource:0}: Error finding container c175c5d68cca3fce2bd6848cac93db854bf7d126ba83861351b14e2aa46b056a: Status 404 returned error can't find the container with id c175c5d68cca3fce2bd6848cac93db854bf7d126ba83861351b14e2aa46b056a Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.206956 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" event={"ID":"539fe2a2-8989-4fb6-a24e-dee67304e96b","Type":"ContainerStarted","Data":"e94c950a2102c2bfa727efdf89f065ed9606ffd6230d59163b476d18227345a9"} Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.210583 4960 generic.go:334] "Generic (PLEG): container finished" podID="e06b264d-7f1e-48b9-a067-1d428f2633f4" containerID="bbc1fe0473e72aad84c3b5b02f3ec95045973f7bb5af1fb4a5270069d2b4f72d" exitCode=0 Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.210654 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9pkx" event={"ID":"e06b264d-7f1e-48b9-a067-1d428f2633f4","Type":"ContainerDied","Data":"bbc1fe0473e72aad84c3b5b02f3ec95045973f7bb5af1fb4a5270069d2b4f72d"} Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.214112 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" event={"ID":"8c352774-9d31-4d0a-a1ee-0a96bfbf6239","Type":"ContainerStarted","Data":"fa679516d0b2bb5dcb8aa1d9d5606655db7e6f7d3e5a2c7d9ad73df04b10950f"} Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.221248 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" event={"ID":"b6b8a229-03cf-4cb0-ab8a-a3133dadfc21","Type":"ContainerStarted","Data":"9aebe4c3ecd591256d380b12992b3955c1e6d38913b12cb420f047e3d15fcd6c"} Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.446237 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f802e6c8-39a5-4acf-b85c-29f235283937-cert\") pod \"openstack-operator-controller-manager-67698bcd47-nz8r6\" (UID: \"f802e6c8-39a5-4acf-b85c-29f235283937\") " pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.453214 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f802e6c8-39a5-4acf-b85c-29f235283937-cert\") pod \"openstack-operator-controller-manager-67698bcd47-nz8r6\" (UID: \"f802e6c8-39a5-4acf-b85c-29f235283937\") " pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.520021 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2"] Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.546514 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d7077dd_5b52_49af_9dd7_c81f361139ec.slice/crio-110389c6ad4838af16e5cadef157763118a138d2266d168c1d46e7e4e536ee27 WatchSource:0}: Error finding container 110389c6ad4838af16e5cadef157763118a138d2266d168c1d46e7e4e536ee27: Status 404 returned error can't find the container with id 110389c6ad4838af16e5cadef157763118a138d2266d168c1d46e7e4e536ee27 Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.556630 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.659457 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.833247 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.842606 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.847440 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.857271 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.862605 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.869107 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.879832 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz"] Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.898299 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32ba83c2_5436_4b71_9a95_e552613c6bff.slice/crio-1c72aa636f1daeeb259ed1a8173ac902a35eb9a6a5bd577c2e54b7e83986ce5c WatchSource:0}: Error finding container 1c72aa636f1daeeb259ed1a8173ac902a35eb9a6a5bd577c2e54b7e83986ce5c: Status 404 returned error can't find the container with id 1c72aa636f1daeeb259ed1a8173ac902a35eb9a6a5bd577c2e54b7e83986ce5c Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.914483 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.924318 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.932347 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5"] Oct 02 07:30:17 crc kubenswrapper[4960]: I1002 07:30:17.938076 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd"] Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.940563 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf515440_a75b_4f32_8c08_2f108da16b56.slice/crio-598984def7d222dfe67edb0dd9029a7877f8f9701d70b453d1a8647c4affd13d WatchSource:0}: Error finding container 598984def7d222dfe67edb0dd9029a7877f8f9701d70b453d1a8647c4affd13d: Status 404 returned error can't find the container with id 598984def7d222dfe67edb0dd9029a7877f8f9701d70b453d1a8647c4affd13d Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.946113 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2af8b0c5_6356_4c12_894e_1bb39e4bf981.slice/crio-723f7239b1faa75af16064f7d475516ecbdd065ff815430a5055d022788b10f9 WatchSource:0}: Error finding container 723f7239b1faa75af16064f7d475516ecbdd065ff815430a5055d022788b10f9: Status 404 returned error can't find the container with id 723f7239b1faa75af16064f7d475516ecbdd065ff815430a5055d022788b10f9 Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.955897 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde18430c_640c_46ac_9eb5_de6216332a80.slice/crio-0fe4c96cf79716338e0a4da45298ccd6e5928521222f5e630f8a433e75550331 WatchSource:0}: Error finding container 0fe4c96cf79716338e0a4da45298ccd6e5928521222f5e630f8a433e75550331: Status 404 returned error can't find the container with id 0fe4c96cf79716338e0a4da45298ccd6e5928521222f5e630f8a433e75550331 Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.957651 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode52d2a9e_b404_44cf_978a_a2fed39d6337.slice/crio-51834a55654d74187dc5807ee9aea3ee52d22b9081fc50de4b117d56e2b0d579 WatchSource:0}: Error finding container 51834a55654d74187dc5807ee9aea3ee52d22b9081fc50de4b117d56e2b0d579: Status 404 returned error can't find the container with id 51834a55654d74187dc5807ee9aea3ee52d22b9081fc50de4b117d56e2b0d579 Oct 02 07:30:17 crc kubenswrapper[4960]: E1002 07:30:17.958757 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ththn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-598c4c8547-fkgnk_openstack-operators(de18430c-640c-46ac-9eb5-de6216332a80): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:30:17 crc kubenswrapper[4960]: W1002 07:30:17.974188 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda786a744_833c_40f9_91a7_8b5f312a626c.slice/crio-7be36ec40d1aa88b34085e7436331570e37f465d0f3bae169fd5625f0655cb3b WatchSource:0}: Error finding container 7be36ec40d1aa88b34085e7436331570e37f465d0f3bae169fd5625f0655cb3b: Status 404 returned error can't find the container with id 7be36ec40d1aa88b34085e7436331570e37f465d0f3bae169fd5625f0655cb3b Oct 02 07:30:17 crc kubenswrapper[4960]: E1002 07:30:17.979492 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n4lck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-cb66d6b59-cmrs5_openstack-operators(73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:30:17 crc kubenswrapper[4960]: E1002 07:30:17.990443 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qzzbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-689b4f76c9-lzlkk_openstack-operators(e52d2a9e-b404-44cf-978a-a2fed39d6337): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.170661 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk"] Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.178988 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g"] Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.188296 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w"] Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.196256 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd"] Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.200932 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4l7sf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-cbdf6dc66-sqb6w_openstack-operators(5529d258-1683-402a-ba4d-fa355405201d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.202340 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-strzj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-68d7bc5569-69k7g_openstack-operators(97605d31-7ce2-4f75-848a-1fcf39e54d3e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.209552 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:fe3439557337d51c30cf1302608e5fc4623b3598ed5c49e9699f13e5abef61cc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:1c99923410d4cd0a721d2cc8a51d91d3ac800d5fda508c972ebe1e85ed2ca4d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:518df53777513ae6af94ff6a267d12981421525b27a5ab5042c2ac95967aa36d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:a5c05d9a996cfc9fdd05454149255f6d72b165369cda4f5cd8f56d4c60245f1a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:8f3b6c2ffde0d9678504683f2a788dafa0834c60f3d325b05bfdae855b0f5f98,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:a05f003cb0c8a5d3ae42d62cb8f06a494a819d801c5e77ebcfb684c9af1238f0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4da48d8f16bfe4f35784b91b85bbd936c35ed26274a3991c54a19e6520c41c3c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:6362514c29a178610a6efabc57c7e05de596cd65279d4ce1faaf7114e3ca9926,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:510e66238e579d31e3fe9e05bf4ea31b1ddc0c1e4a28fe6355a69d0ab9ccac81,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:17a5a8c217d3344d7fbe8dd1b01a4a99194aad59c2683ccd39d077f528c0880e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:c5e9470dcf2202c26ba7112c5aa4803d550367dd8a5f86bbeca6ff8cb404076b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:645d4fca54f09ffba365383ac467cbc0fef04363dd22f9ab2f771721ef557705,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:eac564bfb8a13e2bcd4656ad8e4b45eb57536bd32ed8a33c0945b1fc60414c9d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:dec0378fb73f66dd39f6d7f7bdc6020d9bdf4c57c201eac3c63d7e97667627fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:64d895c5efce47a84c9d4a46a292441d28e5581599ef7cdf8a0a576d2a18288d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:ace365ee06fadfe22f47f3b243d23021ba034d8456c517872db7e4f2965756c5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:010a0936d4dee93552033421bcb26d7ae50bea00f2b59fd3cb2da05991cbb5d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:229d12cd24aabc6b322c3b161c227c8ebc2692f4243b50a6adaa91bb230d30bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:709ea8ccc05ea65e20f9a8eb0ae1020d5524d9ae25123fe8b1f704ad98683bc9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:31f972c87db65413ce2dcf47f25d975f231feda6bedddf33bae9ae2d0eba173e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:b92f7aea6dde34128b0202bce479b3a2d92cf3cbf2a9055c99ece3f995158e03,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:80f9aa236ab23cc06dc2f42f8033b5851fa29bf1e3ee4961f4bf9aec78dc22fd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:453f4e8e8319bbe843e6d78f2c18d98d9ea170ca6541931f5107981cd2a32996,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:5bc1dc22276aecd72a6fd5d7822ceaac25ffaffbbf16e931029fd0c791054dfa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:e992f55a9fbff89bf214ff0c97c73918800c64012443cc76516a361a9986e0d3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:ccd28e86a082f716c528dd19339b20b089babe706b9221d03a483bf5e38eaa85,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:b00e29030c9666484b2f5baac70e3de0a8cab4f58bd3e270f8fbf18953f97bd6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:e2165a561d61825f8c0fb2c0514e8e828f53fa25fe921e79bb0e4b440b10b136,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:b20fa3b0a3a01acd5d506742eaa6500a5e90bb92008ea7fad2e64580cfce9582,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:c669f661c7b636aa2018e5e98d826141cf1626e32230a8985e4cfb6a55e6cb3e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:badaf9205a7dd15f7fa96302c4f1bfd7bbb4c99167ae9f7bf1424e8e7b05d77a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:9dc84b8f1c1db86d71c3d88ae091df6591b72307c2898cf42658b75db6a106a4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:e12f1673e4c4bbd77e1375f7cc9ee883c261542317d1d87f113f6091e90cea29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:85c8d86100270d60e99d2a7ab5e875fa7634a6a8a6c351630fe3b964e1b11f0e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:3b539c6ed6c3a6b66a300e26e3d74432f2297279b11ec05fb31f717ebc5f304e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:a246e5f81db50cd10949057d047bcddd4add5de011c2b5553d82728e845e27d7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:f9f9d3cbaf14d1765dcf3af69c739125ba6b000c490df5f7d1648afe816e1faf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:cd4347fc6be340bb0a5d1fa5046f0372d5a2193d4913ed1e8038b4070fae5169,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:b47d4eae5e865b2f6e1baf1d2bf66aae8e3a91fb6576a1cf9fcb47e0aefd668a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:8e5f7a8e79285c4a6ef668b82b4809210b29a5ace62ee694cad41025c1307086,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:b3bd86ea948427d8b82fcdf709727863679a6152d0df70c9074d02886eb68dd6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:843f85a06093397d9e2dfe9fe89fee5403136ea7ee1ca2fc4f62bcc6424fb560,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:6373beb6be59c9792754c3c754bdbf3323b043f2802c82410fd082e6f9739004,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:05d1262fc9ae9424275f24aa9a18589fb676335ea27a1b438e42d5d18fc15179,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:83e473c5007e76c57b1afa32ad0f72173464e92eaf7a924b198af0c6422a7f28,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:e1961ac4c17954530517f025e7568ddb7b356d20d4298d826e164254da44b823,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:7f9881c69cc601e15662e64fccd927a9b06df2c11f8cafd8ee51d1c2c60446c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:2807588bf47b15de6aa29744ad0e7dd7ee3acc5dbaf9bd008b06546cb976805f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:186c1e750c5d17bff6c8cc78be273be77d846cb5132bdb3d13e4c70ce04ec41c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:1d821861636444ef495e6fbf0f4a767228eefedc4fe51c35ed782687d44fd8fd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0d8e78a9369adac95777be7f2b7ed111f93bc5936b6f284fb1537dc7eae7192a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:87cb6a91c3a9dd4c487e99e9afd38abbca59dcae337f5fec5071c6f92007205d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:a67eff0f7040b234c4d856a076f1df98f3a3d0d2ef02fd499ec580c72c35a57b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:bffa691f6d884b143ff2e1ec18bf26fdfcea39492c68874b12a41aab94fdde38,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:f2074d21cd1928f96272477bf99dfb87ca34a700ddc06754597670b141cf4759,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:b8f132147073a1b2574908075c99ec6ad5cefe6243ae9fe64af2738bd1d01f4e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:c6a0cc0b86366e6436eaa3600e709d9bed8929f94884eeea19be87867ed4e027,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:52498394257731b2e2b61b813434a9cc9b39b2b6bcfaee3c4764ee716807761c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:eb839898ec6bba6e9d552ab64817f49485461b7ebceae1b121002a6e25836ccf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:3f39dceb7272639abe06fe1c218936d8068dfca721c6413bc91868fb77d88720,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:663ff44c4016c5a34460f30ed7f742b8a765ba5cc81897dcf5009d3062637626,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:0e366f865710291a234fb316973854b75946e8ed35dd47353b49de2107ead77c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:a336df8914f0bf065c3f2dd3641c5f96011b53cc504e1a5518979be25bbfc204,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:bf42dfd2e225818662aa28c4bb23204dc47b2b91127ca0e49b085baa1ea7609d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:e034577b9d0781f4548522a99ff2d13f205e67a30d0fc150c7cd0844e6752277,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:a2ac7724a8631b0e30a36b14bfd96711af16439a740cea5fde1405253e9798a0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:f02b83cbbef5604a7387575dad8a299ccb6afa13069dfe3eeec7acd0fda940b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:9229a115d027f5a6576f50870d6fb80d428064662043b3fd7ee3972ee26614c4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:1c50a166bd99e649ad7da7fd5e997b4059e19a6c98a3a8fd4edf3595c15f18c7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:27296ea33566d09733c1122257994f7ab6caf3fd73758c710c81009c12430844,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d88750a3464a2b6dd3bfcfc7222a9579b2aebdf23014ce835ad0b0d8492d0ad9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:aa5f2679b2241034703c548052aa73f1789a0dbc34a87c1cf8e63744f99fac1f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:6d735d97892f6959ead0564ad0d3efc503fe6575f65aed42d7c074a8a99ef08c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:33712c48a2f32c3c7db4d1d77719599fb5619f97d21471ae4b7602be9f5841e2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:2bdfac03b30721b21964bce0b99b45d7f733a3b70104f4afdf4fdc7166aff109,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:4618ac84908528f1203c51515538e5f68bbff19fc3934543d7f89a358a618941,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:bd0ba26707f4b3cc5e98d7a6d6ae05617d47a87f41d89dcbb79a668f9063b963,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:ace1c9bb647fbbac08c2b1e674060962fcba9366b0c6b6bf9c5caad7c80844d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:f0f7b03511d721500f524c1c8a069e9948569b851b91f96c276cae593dd28425,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:37274b985899c529e98703b76a0b8ca6aac1de7a601788d54e90fdc12d1498f0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9g8jw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-787874f5b776jcd_openstack-operators(70d6affd-3067-440a-8b1e-69285828be97): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.258761 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" event={"ID":"73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894","Type":"ContainerStarted","Data":"0ebadb433bfe8645a35c930185de8b053b5577dedb469ebef00ab725ed40f184"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.269266 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" event={"ID":"b9dd3b02-9840-4eef-b378-b7888406a57c","Type":"ContainerStarted","Data":"dfc01c62a78024f47178e6dc1407aef88559621cd38d850dac6e9f0230779edb"} Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.273865 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" podUID="e52d2a9e-b404-44cf-978a-a2fed39d6337" Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.274363 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" podUID="73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894" Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.287678 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" podUID="de18430c-640c-46ac-9eb5-de6216332a80" Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.329150 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6"] Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.356911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9pkx" event={"ID":"e06b264d-7f1e-48b9-a067-1d428f2633f4","Type":"ContainerStarted","Data":"30b8e8995104e48d2786a90f12df46e14c491b476a37f5736fa84547b03389bc"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.356942 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" event={"ID":"97605d31-7ce2-4f75-848a-1fcf39e54d3e","Type":"ContainerStarted","Data":"80e5ac5aa39944e4318eb66242fee726ae32e03652ea0df70f5cf925bb934cb4"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.359053 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" event={"ID":"72b0c3f8-2a66-42da-aa83-298bda8f332d","Type":"ContainerStarted","Data":"996dc570fcbedb3661125d970650770701a448ffc0eb6c7a8cfe93167acc1e51"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.367506 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j9pkx" podStartSLOduration=3.669458363 podStartE2EDuration="11.367489025s" podCreationTimestamp="2025-10-02 07:30:07 +0000 UTC" firstStartedPulling="2025-10-02 07:30:10.031159219 +0000 UTC m=+831.063105546" lastFinishedPulling="2025-10-02 07:30:17.729189921 +0000 UTC m=+838.761136208" observedRunningTime="2025-10-02 07:30:18.366480209 +0000 UTC m=+839.398426496" watchObservedRunningTime="2025-10-02 07:30:18.367489025 +0000 UTC m=+839.399435312" Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.368528 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" event={"ID":"418a02f4-eb9e-4079-969a-c75f8307b078","Type":"ContainerStarted","Data":"87140b6e374bcd518b8097515b02bd38b7f746ac1bdd85d4a31a7924d4f5a0da"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.381008 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" event={"ID":"de18430c-640c-46ac-9eb5-de6216332a80","Type":"ContainerStarted","Data":"0fe4c96cf79716338e0a4da45298ccd6e5928521222f5e630f8a433e75550331"} Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.391915 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" podUID="de18430c-640c-46ac-9eb5-de6216332a80" Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.401805 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" event={"ID":"5529d258-1683-402a-ba4d-fa355405201d","Type":"ContainerStarted","Data":"4e5ed95462f180a6a82e3a5f0bac04ac8794bba44ab007a0eb96865d3e54f770"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.408958 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" event={"ID":"a786a744-833c-40f9-91a7-8b5f312a626c","Type":"ContainerStarted","Data":"7be36ec40d1aa88b34085e7436331570e37f465d0f3bae169fd5625f0655cb3b"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.415658 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" event={"ID":"c0d6bd86-ab92-4f37-be9c-003aab10f910","Type":"ContainerStarted","Data":"f8fa619b7083248acbbb84e944b80a7d4ab6bc60d0b6fea1f58d6cfbd739f382"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.418349 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" event={"ID":"2af8b0c5-6356-4c12-894e-1bb39e4bf981","Type":"ContainerStarted","Data":"723f7239b1faa75af16064f7d475516ecbdd065ff815430a5055d022788b10f9"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.420281 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" event={"ID":"2fac2998-8d33-448f-877a-388b6570a141","Type":"ContainerStarted","Data":"bfb71cc172366b724a4aec33b2731e1eac483388c70dc3ae76b09b7b6e433463"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.422744 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" event={"ID":"e52d2a9e-b404-44cf-978a-a2fed39d6337","Type":"ContainerStarted","Data":"51834a55654d74187dc5807ee9aea3ee52d22b9081fc50de4b117d56e2b0d579"} Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.424172 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" podUID="e52d2a9e-b404-44cf-978a-a2fed39d6337" Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.426580 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" event={"ID":"5d7077dd-5b52-49af-9dd7-c81f361139ec","Type":"ContainerStarted","Data":"110389c6ad4838af16e5cadef157763118a138d2266d168c1d46e7e4e536ee27"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.427725 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" event={"ID":"70d6affd-3067-440a-8b1e-69285828be97","Type":"ContainerStarted","Data":"8a9217bf062935bd58b982997ca177f42b6cd0668d88c044f164fbb09e1b7509"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.431912 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" event={"ID":"cf515440-a75b-4f32-8c08-2f108da16b56","Type":"ContainerStarted","Data":"598984def7d222dfe67edb0dd9029a7877f8f9701d70b453d1a8647c4affd13d"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.435911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" event={"ID":"87d511d4-0cbf-49b2-a2e2-b2e2fc726f60","Type":"ContainerStarted","Data":"40f593175ec8b4199ee5ab865d8869332fe9dc4180fbd5c46c3525a507f099fb"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.437729 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" event={"ID":"6c71ae76-8a1e-4570-8950-8cf4962169e4","Type":"ContainerStarted","Data":"c175c5d68cca3fce2bd6848cac93db854bf7d126ba83861351b14e2aa46b056a"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.444882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" event={"ID":"32ba83c2-5436-4b71-9a95-e552613c6bff","Type":"ContainerStarted","Data":"1c72aa636f1daeeb259ed1a8173ac902a35eb9a6a5bd577c2e54b7e83986ce5c"} Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.451727 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" event={"ID":"4bcfb151-7975-4567-9a1f-0cb7dd128696","Type":"ContainerStarted","Data":"01a98b51a72299d347a207bff27cb1fcb42425d4db79c4f5c5d3c99202687a0c"} Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.463393 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" podUID="5529d258-1683-402a-ba4d-fa355405201d" Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.501578 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" podUID="97605d31-7ce2-4f75-848a-1fcf39e54d3e" Oct 02 07:30:18 crc kubenswrapper[4960]: E1002 07:30:18.512267 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" podUID="70d6affd-3067-440a-8b1e-69285828be97" Oct 02 07:30:18 crc kubenswrapper[4960]: I1002 07:30:18.938225 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx747"] Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.471016 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" event={"ID":"de18430c-640c-46ac-9eb5-de6216332a80","Type":"ContainerStarted","Data":"5b847346378db95f01e8812fed79c57dd8872398f90fe26dfd07f817f93ada9c"} Oct 02 07:30:19 crc kubenswrapper[4960]: E1002 07:30:19.475153 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" podUID="de18430c-640c-46ac-9eb5-de6216332a80" Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.477238 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" event={"ID":"5529d258-1683-402a-ba4d-fa355405201d","Type":"ContainerStarted","Data":"4a62e2cb6aa360185567b4af0271b870f66e1ca04f25291798992462c4c8049c"} Oct 02 07:30:19 crc kubenswrapper[4960]: E1002 07:30:19.479867 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" podUID="5529d258-1683-402a-ba4d-fa355405201d" Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.483569 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" event={"ID":"73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894","Type":"ContainerStarted","Data":"4f7f89128160fb547d7184a0e0f16450b0b6cbe08cc8a3895d112bd75ff7c0bc"} Oct 02 07:30:19 crc kubenswrapper[4960]: E1002 07:30:19.487378 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" podUID="73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894" Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.488388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" event={"ID":"e52d2a9e-b404-44cf-978a-a2fed39d6337","Type":"ContainerStarted","Data":"6b62a4d3678ae32146c52e3d0a43b37ea50cc41718ef642f6910bd9afdc36409"} Oct 02 07:30:19 crc kubenswrapper[4960]: E1002 07:30:19.502178 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" podUID="e52d2a9e-b404-44cf-978a-a2fed39d6337" Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.510480 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" event={"ID":"f802e6c8-39a5-4acf-b85c-29f235283937","Type":"ContainerStarted","Data":"6ef5fef55c11b60e4517b29f76f0f05aa7fa7d908bfea738ea164de274c9856a"} Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.510540 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" event={"ID":"f802e6c8-39a5-4acf-b85c-29f235283937","Type":"ContainerStarted","Data":"bf02818836990123c4be1068a07160c1faaa3f02306a48272df6d2811ecb5c45"} Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.510550 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" event={"ID":"f802e6c8-39a5-4acf-b85c-29f235283937","Type":"ContainerStarted","Data":"a9a8a9bd02593c53757911c610b44deb3557b52010d80dc822b6571055a90b71"} Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.511441 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.527659 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" event={"ID":"70d6affd-3067-440a-8b1e-69285828be97","Type":"ContainerStarted","Data":"fd3d829225a9231212a0d054a3132c0baf146e54bf15c3bb5319f9a66f65b3ee"} Oct 02 07:30:19 crc kubenswrapper[4960]: E1002 07:30:19.531316 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" podUID="70d6affd-3067-440a-8b1e-69285828be97" Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.551529 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" event={"ID":"97605d31-7ce2-4f75-848a-1fcf39e54d3e","Type":"ContainerStarted","Data":"4ccb895cf16f11b0556b3aefe2063ded8ee31e3674fb001a944a8ce5a2e9f31e"} Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.551712 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kx747" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="registry-server" containerID="cri-o://05fa352a898b8ce0074aa520b060a8046651106341e98697dc9b7615cd9de27e" gracePeriod=2 Oct 02 07:30:19 crc kubenswrapper[4960]: E1002 07:30:19.556195 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" podUID="97605d31-7ce2-4f75-848a-1fcf39e54d3e" Oct 02 07:30:19 crc kubenswrapper[4960]: I1002 07:30:19.647300 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" podStartSLOduration=3.647263509 podStartE2EDuration="3.647263509s" podCreationTimestamp="2025-10-02 07:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:30:19.643097257 +0000 UTC m=+840.675043554" watchObservedRunningTime="2025-10-02 07:30:19.647263509 +0000 UTC m=+840.679209796" Oct 02 07:30:20 crc kubenswrapper[4960]: I1002 07:30:20.577521 4960 generic.go:334] "Generic (PLEG): container finished" podID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerID="05fa352a898b8ce0074aa520b060a8046651106341e98697dc9b7615cd9de27e" exitCode=0 Oct 02 07:30:20 crc kubenswrapper[4960]: I1002 07:30:20.577573 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx747" event={"ID":"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9","Type":"ContainerDied","Data":"05fa352a898b8ce0074aa520b060a8046651106341e98697dc9b7615cd9de27e"} Oct 02 07:30:20 crc kubenswrapper[4960]: E1002 07:30:20.582327 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" podUID="97605d31-7ce2-4f75-848a-1fcf39e54d3e" Oct 02 07:30:20 crc kubenswrapper[4960]: E1002 07:30:20.584864 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" podUID="73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894" Oct 02 07:30:20 crc kubenswrapper[4960]: E1002 07:30:20.584942 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" podUID="e52d2a9e-b404-44cf-978a-a2fed39d6337" Oct 02 07:30:20 crc kubenswrapper[4960]: E1002 07:30:20.585014 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" podUID="5529d258-1683-402a-ba4d-fa355405201d" Oct 02 07:30:20 crc kubenswrapper[4960]: E1002 07:30:20.590873 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" podUID="70d6affd-3067-440a-8b1e-69285828be97" Oct 02 07:30:20 crc kubenswrapper[4960]: E1002 07:30:20.601944 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" podUID="de18430c-640c-46ac-9eb5-de6216332a80" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.143924 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.214670 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rtsp\" (UniqueName: \"kubernetes.io/projected/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-kube-api-access-6rtsp\") pod \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.214758 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-utilities\") pod \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.214890 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-catalog-content\") pod \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\" (UID: \"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9\") " Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.217205 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-utilities" (OuterVolumeSpecName: "utilities") pod "c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" (UID: "c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.222880 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-kube-api-access-6rtsp" (OuterVolumeSpecName: "kube-api-access-6rtsp") pod "c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" (UID: "c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9"). InnerVolumeSpecName "kube-api-access-6rtsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.228719 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" (UID: "c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.317446 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rtsp\" (UniqueName: \"kubernetes.io/projected/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-kube-api-access-6rtsp\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.317513 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.317531 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.591512 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx747" event={"ID":"c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9","Type":"ContainerDied","Data":"17b82905f44b5fa83c1b4f9c977c00f31ebf8ffb8a7307ffdb18c077b7308047"} Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.591549 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx747" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.591577 4960 scope.go:117] "RemoveContainer" containerID="05fa352a898b8ce0074aa520b060a8046651106341e98697dc9b7615cd9de27e" Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.625968 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx747"] Oct 02 07:30:21 crc kubenswrapper[4960]: I1002 07:30:21.632312 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx747"] Oct 02 07:30:22 crc kubenswrapper[4960]: I1002 07:30:22.339876 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" path="/var/lib/kubelet/pods/c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9/volumes" Oct 02 07:30:27 crc kubenswrapper[4960]: I1002 07:30:27.665754 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-67698bcd47-nz8r6" Oct 02 07:30:28 crc kubenswrapper[4960]: I1002 07:30:28.271834 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:28 crc kubenswrapper[4960]: I1002 07:30:28.271954 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:28 crc kubenswrapper[4960]: I1002 07:30:28.349143 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:28 crc kubenswrapper[4960]: I1002 07:30:28.697912 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j9pkx" Oct 02 07:30:28 crc kubenswrapper[4960]: I1002 07:30:28.770140 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9pkx"] Oct 02 07:30:28 crc kubenswrapper[4960]: I1002 07:30:28.810182 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9z79"] Oct 02 07:30:28 crc kubenswrapper[4960]: I1002 07:30:28.810634 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c9z79" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="registry-server" containerID="cri-o://e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" gracePeriod=2 Oct 02 07:30:29 crc kubenswrapper[4960]: E1002 07:30:29.324927 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:30:29 crc kubenswrapper[4960]: E1002 07:30:29.325387 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:30:29 crc kubenswrapper[4960]: E1002 07:30:29.325891 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:30:29 crc kubenswrapper[4960]: E1002 07:30:29.325930 4960 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-c9z79" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="registry-server" Oct 02 07:30:29 crc kubenswrapper[4960]: I1002 07:30:29.670675 4960 generic.go:334] "Generic (PLEG): container finished" podID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" exitCode=0 Oct 02 07:30:29 crc kubenswrapper[4960]: I1002 07:30:29.670769 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9z79" event={"ID":"7c4be745-c2cd-4a2b-844b-67572c49ca08","Type":"ContainerDied","Data":"e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be"} Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.029823 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-drs8j"] Oct 02 07:30:33 crc kubenswrapper[4960]: E1002 07:30:33.031542 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="extract-utilities" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.031619 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="extract-utilities" Oct 02 07:30:33 crc kubenswrapper[4960]: E1002 07:30:33.031711 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="extract-content" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.031769 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="extract-content" Oct 02 07:30:33 crc kubenswrapper[4960]: E1002 07:30:33.031829 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="registry-server" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.031883 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="registry-server" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.032099 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1fba335-8b2a-4f7c-a78f-6e5b3a2cfee9" containerName="registry-server" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.033184 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.045384 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-drs8j"] Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.120761 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-catalog-content\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.120917 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-utilities\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.121016 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnll5\" (UniqueName: \"kubernetes.io/projected/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-kube-api-access-xnll5\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.222104 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-utilities\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.222185 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnll5\" (UniqueName: \"kubernetes.io/projected/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-kube-api-access-xnll5\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.222229 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-catalog-content\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.222583 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-utilities\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.222741 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-catalog-content\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.243738 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnll5\" (UniqueName: \"kubernetes.io/projected/3a353733-ab76-4e1f-8c4d-9c4dd542e0e9-kube-api-access-xnll5\") pod \"certified-operators-drs8j\" (UID: \"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9\") " pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:33 crc kubenswrapper[4960]: I1002 07:30:33.363503 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:36 crc kubenswrapper[4960]: E1002 07:30:36.145239 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302" Oct 02 07:30:36 crc kubenswrapper[4960]: E1002 07:30:36.147544 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kh8q6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-84c745747f-lk5kd_openstack-operators(a786a744-833c-40f9-91a7-8b5f312a626c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:30:38 crc kubenswrapper[4960]: E1002 07:30:38.977699 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3" Oct 02 07:30:38 crc kubenswrapper[4960]: E1002 07:30:38.978344 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nflq2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-5c8fdc4d5c-hbzp7_openstack-operators(cf515440-a75b-4f32-8c08-2f108da16b56): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.325826 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.326136 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.326581 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.326706 4960 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-c9z79" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="registry-server" Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.405223 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397" Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.405413 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tvjm6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-679b4759bb-cd9cz_openstack-operators(72b0c3f8-2a66-42da-aa83-298bda8f332d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.503328 4960 scope.go:117] "RemoveContainer" containerID="f4d03ef8741b23acb29c81dc03b5462f69d601ef7e8eb49aaf931265c274411d" Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.586853 4960 scope.go:117] "RemoveContainer" containerID="a10b75b9fb1be64790ab2dbfb0b0d7f06b456e1eb7bf90b420146b5292353845" Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.772568 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" podUID="cf515440-a75b-4f32-8c08-2f108da16b56" Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.775050 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" podUID="a786a744-833c-40f9-91a7-8b5f312a626c" Oct 02 07:30:39 crc kubenswrapper[4960]: E1002 07:30:39.846301 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" podUID="72b0c3f8-2a66-42da-aa83-298bda8f332d" Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.874085 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.946695 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-utilities\") pod \"7c4be745-c2cd-4a2b-844b-67572c49ca08\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.946751 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr57p\" (UniqueName: \"kubernetes.io/projected/7c4be745-c2cd-4a2b-844b-67572c49ca08-kube-api-access-tr57p\") pod \"7c4be745-c2cd-4a2b-844b-67572c49ca08\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.947093 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-catalog-content\") pod \"7c4be745-c2cd-4a2b-844b-67572c49ca08\" (UID: \"7c4be745-c2cd-4a2b-844b-67572c49ca08\") " Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.948202 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-utilities" (OuterVolumeSpecName: "utilities") pod "7c4be745-c2cd-4a2b-844b-67572c49ca08" (UID: "7c4be745-c2cd-4a2b-844b-67572c49ca08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:39 crc kubenswrapper[4960]: I1002 07:30:39.956815 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4be745-c2cd-4a2b-844b-67572c49ca08-kube-api-access-tr57p" (OuterVolumeSpecName: "kube-api-access-tr57p") pod "7c4be745-c2cd-4a2b-844b-67572c49ca08" (UID: "7c4be745-c2cd-4a2b-844b-67572c49ca08"). InnerVolumeSpecName "kube-api-access-tr57p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.008310 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c4be745-c2cd-4a2b-844b-67572c49ca08" (UID: "7c4be745-c2cd-4a2b-844b-67572c49ca08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.027913 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-drs8j"] Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.048717 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.048743 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4be745-c2cd-4a2b-844b-67572c49ca08-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.048754 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr57p\" (UniqueName: \"kubernetes.io/projected/7c4be745-c2cd-4a2b-844b-67572c49ca08-kube-api-access-tr57p\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.889774 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9z79" event={"ID":"7c4be745-c2cd-4a2b-844b-67572c49ca08","Type":"ContainerDied","Data":"6abcb5901e984ed810df9436b682ccd18f7f134e21014040e7e2ff9ccc2d6711"} Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.890195 4960 scope.go:117] "RemoveContainer" containerID="e3e8cf4fbbc29f1b369cca8ec6c1049ede62daab0ad0d7668d866e4c39c275be" Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.890333 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9z79" Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.924473 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" event={"ID":"32ba83c2-5436-4b71-9a95-e552613c6bff","Type":"ContainerStarted","Data":"c59dc9eb37aaee373dcf101faf8d434dd1eb370220112fcf31ca8a67542ace42"} Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.950595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" event={"ID":"539fe2a2-8989-4fb6-a24e-dee67304e96b","Type":"ContainerStarted","Data":"2d105e93616c8e1ff6a0d160bf4e5935d95b3af6f936d39852533270da40ba1b"} Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.956998 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" event={"ID":"5d7077dd-5b52-49af-9dd7-c81f361139ec","Type":"ContainerStarted","Data":"2742b8fa5bc8aa5d637de907b35d811599e7e807d369856d4d1d6f4c8620d9d4"} Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.965526 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" event={"ID":"72b0c3f8-2a66-42da-aa83-298bda8f332d","Type":"ContainerStarted","Data":"0933c4507d88f8b9ad5e450c889204022f49c00043140a219012c9c0604a65cf"} Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.982655 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" event={"ID":"418a02f4-eb9e-4079-969a-c75f8307b078","Type":"ContainerStarted","Data":"803a7433c8cdb19ab32fc77b3c9496d9c8a9fd39edece0d0e8b7c53735adcd5c"} Oct 02 07:30:40 crc kubenswrapper[4960]: I1002 07:30:40.992955 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" event={"ID":"6c71ae76-8a1e-4570-8950-8cf4962169e4","Type":"ContainerStarted","Data":"41dad70a745abbd0a98e24c783b8b73a4605cd7541f6ec4180c5d883d9435e37"} Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.005116 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drs8j" event={"ID":"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9","Type":"ContainerStarted","Data":"0daefb9bebf71c6b2c0165b79545e81f0a492ba3377123872ec311efe16cccc8"} Oct 02 07:30:41 crc kubenswrapper[4960]: E1002 07:30:41.006056 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" podUID="72b0c3f8-2a66-42da-aa83-298bda8f332d" Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.034056 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9z79"] Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.040129 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" event={"ID":"b9dd3b02-9840-4eef-b378-b7888406a57c","Type":"ContainerStarted","Data":"5128235c03b1822f831167bf5c353633f8765ffb29cfe2b9200cf3b0aef9a0fb"} Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.058615 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c9z79"] Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.059025 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" event={"ID":"cf515440-a75b-4f32-8c08-2f108da16b56","Type":"ContainerStarted","Data":"50cd78918262742199556ab481e111255e61bdb8bdc822b765611f8a44f2c8ab"} Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.076266 4960 scope.go:117] "RemoveContainer" containerID="80561f5224e962341e7ba53209b2d22283980c55363a175c43837b274185f8db" Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.090436 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" event={"ID":"87d511d4-0cbf-49b2-a2e2-b2e2fc726f60","Type":"ContainerStarted","Data":"80764720f909913bb9021def690d0db800d11f8e364e54136960c26f939f36d7"} Oct 02 07:30:41 crc kubenswrapper[4960]: E1002 07:30:41.119130 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" podUID="cf515440-a75b-4f32-8c08-2f108da16b56" Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.131027 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" event={"ID":"b6b8a229-03cf-4cb0-ab8a-a3133dadfc21","Type":"ContainerStarted","Data":"45de15896429b61ce09ce5e27a1f485c908d238c515fcbbbfcf3da1ceccf9d13"} Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.139757 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" event={"ID":"2fac2998-8d33-448f-877a-388b6570a141","Type":"ContainerStarted","Data":"031ab2758c0063afc458b43a40f04ab722adb180a30a03b4e3e9de4c240adcee"} Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.154796 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" event={"ID":"a786a744-833c-40f9-91a7-8b5f312a626c","Type":"ContainerStarted","Data":"c960a86271042b9abd8b71bdf2171c63367ca8781d1e4e77345f302d7486a606"} Oct 02 07:30:41 crc kubenswrapper[4960]: E1002 07:30:41.161220 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" podUID="a786a744-833c-40f9-91a7-8b5f312a626c" Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.181530 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk" podStartSLOduration=3.8697457870000003 podStartE2EDuration="25.181507274s" podCreationTimestamp="2025-10-02 07:30:16 +0000 UTC" firstStartedPulling="2025-10-02 07:30:18.195489169 +0000 UTC m=+839.227435456" lastFinishedPulling="2025-10-02 07:30:39.507250656 +0000 UTC m=+860.539196943" observedRunningTime="2025-10-02 07:30:41.180065324 +0000 UTC m=+862.212011611" watchObservedRunningTime="2025-10-02 07:30:41.181507274 +0000 UTC m=+862.213453561" Oct 02 07:30:41 crc kubenswrapper[4960]: I1002 07:30:41.241375 4960 scope.go:117] "RemoveContainer" containerID="f890a7397a96b6828e357e8a557dc2c9c20b1ed017cb37b36c03705f02910170" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.170566 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" event={"ID":"4bcfb151-7975-4567-9a1f-0cb7dd128696","Type":"ContainerStarted","Data":"7294160629da7965a3c2ff61e44c18f672cbb5493e3cad746a63153158056e2b"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.172502 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" event={"ID":"5d7077dd-5b52-49af-9dd7-c81f361139ec","Type":"ContainerStarted","Data":"0f9e5ea72cb2f49bdac86e75b2765c2241c353efacf0cbae55c7317efa6a374d"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.172578 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.179196 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" event={"ID":"32ba83c2-5436-4b71-9a95-e552613c6bff","Type":"ContainerStarted","Data":"6e772b5d241de2aef704ad374a4073570aa72d1967112679349001dd4c475425"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.179325 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.182661 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" event={"ID":"8c352774-9d31-4d0a-a1ee-0a96bfbf6239","Type":"ContainerStarted","Data":"ae7a3de30c37befeace8b65ba77cdf916aa2c1819a13b2643482c2710d3f8701"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.182762 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" event={"ID":"8c352774-9d31-4d0a-a1ee-0a96bfbf6239","Type":"ContainerStarted","Data":"26355618fb45bf8fa57f8a6262361e3010f66173eb1d52aa0e6702af16275ae3"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.182797 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.189288 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" event={"ID":"c0d6bd86-ab92-4f37-be9c-003aab10f910","Type":"ContainerStarted","Data":"7b61a1cbae567bf2e993248aba14c0acf1d11091627027c7a4a9c32731475281"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.191637 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" event={"ID":"87d511d4-0cbf-49b2-a2e2-b2e2fc726f60","Type":"ContainerStarted","Data":"2a38b9340767d94cbcf44bfc1faceb31fbc0d51a9c379a29810d3ed0e85e3231"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.192374 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.192578 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" podStartSLOduration=5.238766966 podStartE2EDuration="27.192562367s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.553405764 +0000 UTC m=+838.585352051" lastFinishedPulling="2025-10-02 07:30:39.507201165 +0000 UTC m=+860.539147452" observedRunningTime="2025-10-02 07:30:42.189552845 +0000 UTC m=+863.221499132" watchObservedRunningTime="2025-10-02 07:30:42.192562367 +0000 UTC m=+863.224508654" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.201621 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" event={"ID":"418a02f4-eb9e-4079-969a-c75f8307b078","Type":"ContainerStarted","Data":"8bba57498c1a1cfca2022ba65af628530c4caef947e5ea1c3c1bd0924aa6eb8b"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.203324 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.212452 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" podStartSLOduration=4.761338224 podStartE2EDuration="27.212408023s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.043365485 +0000 UTC m=+838.075311772" lastFinishedPulling="2025-10-02 07:30:39.494435284 +0000 UTC m=+860.526381571" observedRunningTime="2025-10-02 07:30:42.206008797 +0000 UTC m=+863.237955094" watchObservedRunningTime="2025-10-02 07:30:42.212408023 +0000 UTC m=+863.244354310" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.222604 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" event={"ID":"6c71ae76-8a1e-4570-8950-8cf4962169e4","Type":"ContainerStarted","Data":"a648c537604297797e317d191d547392c79653490e1add1ceba0782ef7172f85"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.223642 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.232819 4960 generic.go:334] "Generic (PLEG): container finished" podID="3a353733-ab76-4e1f-8c4d-9c4dd542e0e9" containerID="808d6724e2e0e1db440530f689756f12551316a4d5bc4a522e38f028aaddd68f" exitCode=0 Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.233078 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drs8j" event={"ID":"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9","Type":"ContainerDied","Data":"808d6724e2e0e1db440530f689756f12551316a4d5bc4a522e38f028aaddd68f"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.237611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" event={"ID":"b9dd3b02-9840-4eef-b378-b7888406a57c","Type":"ContainerStarted","Data":"b7934945bdc798f90201b35126da335ef6b1e2bd039e084748c37c2941a7f4bc"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.238151 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.240214 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" podStartSLOduration=5.640408864 podStartE2EDuration="27.240184594s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.907873287 +0000 UTC m=+838.939819574" lastFinishedPulling="2025-10-02 07:30:39.507649017 +0000 UTC m=+860.539595304" observedRunningTime="2025-10-02 07:30:42.232101413 +0000 UTC m=+863.264047710" watchObservedRunningTime="2025-10-02 07:30:42.240184594 +0000 UTC m=+863.272130881" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.241789 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" event={"ID":"539fe2a2-8989-4fb6-a24e-dee67304e96b","Type":"ContainerStarted","Data":"a5bb57e8466bb4137114062ae06dd5493a66f0e4a5fb415d47560bb42e7d0aea"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.242074 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.249617 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" event={"ID":"2af8b0c5-6356-4c12-894e-1bb39e4bf981","Type":"ContainerStarted","Data":"3fc3bd744ee21627555b00ab9727145eac242e50dc678a03cdee50166fb637ea"} Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.257110 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" event={"ID":"b6b8a229-03cf-4cb0-ab8a-a3133dadfc21","Type":"ContainerStarted","Data":"9259c93aa736630c73b4dd7e6fe17abbc8aea9c6052d39f785a23b0817e385ec"} Oct 02 07:30:42 crc kubenswrapper[4960]: E1002 07:30:42.258741 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" podUID="a786a744-833c-40f9-91a7-8b5f312a626c" Oct 02 07:30:42 crc kubenswrapper[4960]: E1002 07:30:42.265816 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" podUID="72b0c3f8-2a66-42da-aa83-298bda8f332d" Oct 02 07:30:42 crc kubenswrapper[4960]: E1002 07:30:42.266005 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" podUID="cf515440-a75b-4f32-8c08-2f108da16b56" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.273992 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" podStartSLOduration=4.9802957679999995 podStartE2EDuration="27.27280986s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.213640754 +0000 UTC m=+838.245587041" lastFinishedPulling="2025-10-02 07:30:39.506154846 +0000 UTC m=+860.538101133" observedRunningTime="2025-10-02 07:30:42.26550021 +0000 UTC m=+863.297446487" watchObservedRunningTime="2025-10-02 07:30:42.27280986 +0000 UTC m=+863.304756137" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.287766 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" podStartSLOduration=5.6673777 podStartE2EDuration="27.28774648s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.898637369 +0000 UTC m=+838.930583646" lastFinishedPulling="2025-10-02 07:30:39.519006139 +0000 UTC m=+860.550952426" observedRunningTime="2025-10-02 07:30:42.284026788 +0000 UTC m=+863.315973075" watchObservedRunningTime="2025-10-02 07:30:42.28774648 +0000 UTC m=+863.319692767" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.322134 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" podStartSLOduration=5.658201266 podStartE2EDuration="27.322104503s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.858364281 +0000 UTC m=+838.890310568" lastFinishedPulling="2025-10-02 07:30:39.522267518 +0000 UTC m=+860.554213805" observedRunningTime="2025-10-02 07:30:42.301988312 +0000 UTC m=+863.333934599" watchObservedRunningTime="2025-10-02 07:30:42.322104503 +0000 UTC m=+863.354050790" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.329280 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" podStartSLOduration=4.703521793 podStartE2EDuration="27.32926067s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:16.881710815 +0000 UTC m=+837.913657092" lastFinishedPulling="2025-10-02 07:30:39.507449682 +0000 UTC m=+860.539395969" observedRunningTime="2025-10-02 07:30:42.32272897 +0000 UTC m=+863.354675257" watchObservedRunningTime="2025-10-02 07:30:42.32926067 +0000 UTC m=+863.361206957" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.350946 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" path="/var/lib/kubelet/pods/7c4be745-c2cd-4a2b-844b-67572c49ca08/volumes" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.379266 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" podStartSLOduration=5.427359024 podStartE2EDuration="27.379245292s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.553222319 +0000 UTC m=+838.585168616" lastFinishedPulling="2025-10-02 07:30:39.505108597 +0000 UTC m=+860.537054884" observedRunningTime="2025-10-02 07:30:42.378860611 +0000 UTC m=+863.410806898" watchObservedRunningTime="2025-10-02 07:30:42.379245292 +0000 UTC m=+863.411191579" Oct 02 07:30:42 crc kubenswrapper[4960]: I1002 07:30:42.401880 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" podStartSLOduration=4.656104062 podStartE2EDuration="27.40177637s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:16.761106315 +0000 UTC m=+837.793052602" lastFinishedPulling="2025-10-02 07:30:39.506778613 +0000 UTC m=+860.538724910" observedRunningTime="2025-10-02 07:30:42.396027983 +0000 UTC m=+863.427974270" watchObservedRunningTime="2025-10-02 07:30:42.40177637 +0000 UTC m=+863.433722647" Oct 02 07:30:43 crc kubenswrapper[4960]: I1002 07:30:43.277696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" event={"ID":"c0d6bd86-ab92-4f37-be9c-003aab10f910","Type":"ContainerStarted","Data":"10c7ea6dcbfdb0c55f38b39595e8d93835e10dbdbfc5e0a87e124a33b95a6259"} Oct 02 07:30:43 crc kubenswrapper[4960]: I1002 07:30:43.278176 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" Oct 02 07:30:43 crc kubenswrapper[4960]: I1002 07:30:43.285949 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" event={"ID":"4bcfb151-7975-4567-9a1f-0cb7dd128696","Type":"ContainerStarted","Data":"f89648051af1ecf83db8a53365257e5e716e3880bcaa367989d05e69f147628f"} Oct 02 07:30:43 crc kubenswrapper[4960]: I1002 07:30:43.286009 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" Oct 02 07:30:43 crc kubenswrapper[4960]: I1002 07:30:43.287022 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" Oct 02 07:30:43 crc kubenswrapper[4960]: I1002 07:30:43.326602 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" podStartSLOduration=6.731717716 podStartE2EDuration="28.326579616s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.898430383 +0000 UTC m=+838.930376670" lastFinishedPulling="2025-10-02 07:30:39.493292283 +0000 UTC m=+860.525238570" observedRunningTime="2025-10-02 07:30:43.300286805 +0000 UTC m=+864.332233092" watchObservedRunningTime="2025-10-02 07:30:43.326579616 +0000 UTC m=+864.358525903" Oct 02 07:30:43 crc kubenswrapper[4960]: I1002 07:30:43.327894 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" podStartSLOduration=6.029983876 podStartE2EDuration="28.327887242s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.236211118 +0000 UTC m=+838.268157405" lastFinishedPulling="2025-10-02 07:30:39.534114484 +0000 UTC m=+860.566060771" observedRunningTime="2025-10-02 07:30:43.31977564 +0000 UTC m=+864.351721927" watchObservedRunningTime="2025-10-02 07:30:43.327887242 +0000 UTC m=+864.359833529" Oct 02 07:30:45 crc kubenswrapper[4960]: I1002 07:30:45.793512 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-mfg89" Oct 02 07:30:45 crc kubenswrapper[4960]: I1002 07:30:45.815413 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-dfpnm" Oct 02 07:30:45 crc kubenswrapper[4960]: I1002 07:30:45.839899 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-7qjq7" Oct 02 07:30:45 crc kubenswrapper[4960]: I1002 07:30:45.867969 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-5mk92" Oct 02 07:30:45 crc kubenswrapper[4960]: I1002 07:30:45.941446 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-wm4wc" Oct 02 07:30:46 crc kubenswrapper[4960]: I1002 07:30:46.065766 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-54crk" Oct 02 07:30:46 crc kubenswrapper[4960]: I1002 07:30:46.181143 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-g9bwh" Oct 02 07:30:46 crc kubenswrapper[4960]: I1002 07:30:46.215942 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-kgzd2" Oct 02 07:30:46 crc kubenswrapper[4960]: I1002 07:30:46.340595 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-6lpjs" Oct 02 07:30:46 crc kubenswrapper[4960]: I1002 07:30:46.382283 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-4vtmz" Oct 02 07:30:46 crc kubenswrapper[4960]: I1002 07:30:46.508037 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xhd8b" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.338882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" event={"ID":"73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894","Type":"ContainerStarted","Data":"b1d5c9e899fd50a5bcd94706ad00080150311b4499e5c60b594742d74825b679"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.339453 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" event={"ID":"e52d2a9e-b404-44cf-978a-a2fed39d6337","Type":"ContainerStarted","Data":"fab60e66a16e01e8bbcb91db4529e45b3575f9286f4c8e9703efa32c2ffd67f9"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.339497 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" event={"ID":"70d6affd-3067-440a-8b1e-69285828be97","Type":"ContainerStarted","Data":"fc35b29f7771087d3bea424872a06a646d6faffbe808b6918f655bf6ea9eba2c"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.340018 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.341673 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" event={"ID":"97605d31-7ce2-4f75-848a-1fcf39e54d3e","Type":"ContainerStarted","Data":"67b6b409931d5af37bf9feb04e0ef09dc4feda727eca5a95d9d29f994aa4d6a7"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.341920 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.344267 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" event={"ID":"2af8b0c5-6356-4c12-894e-1bb39e4bf981","Type":"ContainerStarted","Data":"b5f946379807bd056bfbe48da5c3692be3cfca6a24df7d3cd7ed949144f0df5d"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.344483 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.348752 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" event={"ID":"de18430c-640c-46ac-9eb5-de6216332a80","Type":"ContainerStarted","Data":"b045907995b0e5f6e18288636e9e2015ef3c69109449d4701454a382c37d7539"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.348969 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.349108 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.354412 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" event={"ID":"5529d258-1683-402a-ba4d-fa355405201d","Type":"ContainerStarted","Data":"35b92ac4bf724f020ead22c4ceda7e25d661fbd0079bb6e60143c399455bd99b"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.354595 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.357236 4960 generic.go:334] "Generic (PLEG): container finished" podID="3a353733-ab76-4e1f-8c4d-9c4dd542e0e9" containerID="de798715d75987d1dd4b5d96b239fc2eaa40535fd5e283b392528f205795520a" exitCode=0 Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.357302 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drs8j" event={"ID":"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9","Type":"ContainerDied","Data":"de798715d75987d1dd4b5d96b239fc2eaa40535fd5e283b392528f205795520a"} Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.367415 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" podStartSLOduration=3.897986821 podStartE2EDuration="33.367391326s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.990207472 +0000 UTC m=+839.022153759" lastFinishedPulling="2025-10-02 07:30:47.459611957 +0000 UTC m=+868.491558264" observedRunningTime="2025-10-02 07:30:48.363727805 +0000 UTC m=+869.395674082" watchObservedRunningTime="2025-10-02 07:30:48.367391326 +0000 UTC m=+869.399337633" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.382448 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" podStartSLOduration=3.908963628 podStartE2EDuration="33.382427478s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.979325 +0000 UTC m=+839.011271287" lastFinishedPulling="2025-10-02 07:30:47.45278884 +0000 UTC m=+868.484735137" observedRunningTime="2025-10-02 07:30:48.376177467 +0000 UTC m=+869.408123754" watchObservedRunningTime="2025-10-02 07:30:48.382427478 +0000 UTC m=+869.414373785" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.398312 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-hkswg" podStartSLOduration=11.756207924 podStartE2EDuration="33.398294354s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.950098207 +0000 UTC m=+838.982044494" lastFinishedPulling="2025-10-02 07:30:39.592184637 +0000 UTC m=+860.624130924" observedRunningTime="2025-10-02 07:30:48.395725494 +0000 UTC m=+869.427671791" watchObservedRunningTime="2025-10-02 07:30:48.398294354 +0000 UTC m=+869.430240641" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.481878 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" podStartSLOduration=4.237855122 podStartE2EDuration="33.481834187s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:18.209078523 +0000 UTC m=+839.241024810" lastFinishedPulling="2025-10-02 07:30:47.453057588 +0000 UTC m=+868.485003875" observedRunningTime="2025-10-02 07:30:48.472492661 +0000 UTC m=+869.504438958" watchObservedRunningTime="2025-10-02 07:30:48.481834187 +0000 UTC m=+869.513780464" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.501708 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" podStartSLOduration=3.2505529060000002 podStartE2EDuration="32.501686262s" podCreationTimestamp="2025-10-02 07:30:16 +0000 UTC" firstStartedPulling="2025-10-02 07:30:18.202190279 +0000 UTC m=+839.234136566" lastFinishedPulling="2025-10-02 07:30:47.453323635 +0000 UTC m=+868.485269922" observedRunningTime="2025-10-02 07:30:48.500613983 +0000 UTC m=+869.532560280" watchObservedRunningTime="2025-10-02 07:30:48.501686262 +0000 UTC m=+869.533632549" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.527393 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" podStartSLOduration=4.266565506 podStartE2EDuration="33.527371137s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:18.20076677 +0000 UTC m=+839.232713057" lastFinishedPulling="2025-10-02 07:30:47.461572401 +0000 UTC m=+868.493518688" observedRunningTime="2025-10-02 07:30:48.521780524 +0000 UTC m=+869.553726811" watchObservedRunningTime="2025-10-02 07:30:48.527371137 +0000 UTC m=+869.559317424" Oct 02 07:30:48 crc kubenswrapper[4960]: I1002 07:30:48.551396 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" podStartSLOduration=4.05096994 podStartE2EDuration="33.551378516s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.958539723 +0000 UTC m=+838.990486010" lastFinishedPulling="2025-10-02 07:30:47.458948299 +0000 UTC m=+868.490894586" observedRunningTime="2025-10-02 07:30:48.547995454 +0000 UTC m=+869.579941741" watchObservedRunningTime="2025-10-02 07:30:48.551378516 +0000 UTC m=+869.583324803" Oct 02 07:30:49 crc kubenswrapper[4960]: I1002 07:30:49.370206 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drs8j" event={"ID":"3a353733-ab76-4e1f-8c4d-9c4dd542e0e9","Type":"ContainerStarted","Data":"29a28999310859242a0857b4be18446e2423e06d3b54a4c124123fb279720c24"} Oct 02 07:30:49 crc kubenswrapper[4960]: I1002 07:30:49.405633 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-drs8j" podStartSLOduration=10.171370496 podStartE2EDuration="16.405611765s" podCreationTimestamp="2025-10-02 07:30:33 +0000 UTC" firstStartedPulling="2025-10-02 07:30:42.629270305 +0000 UTC m=+863.661216592" lastFinishedPulling="2025-10-02 07:30:48.863511574 +0000 UTC m=+869.895457861" observedRunningTime="2025-10-02 07:30:49.400555925 +0000 UTC m=+870.432502222" watchObservedRunningTime="2025-10-02 07:30:49.405611765 +0000 UTC m=+870.437558052" Oct 02 07:30:53 crc kubenswrapper[4960]: I1002 07:30:53.364198 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:53 crc kubenswrapper[4960]: I1002 07:30:53.364667 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:53 crc kubenswrapper[4960]: I1002 07:30:53.441955 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:30:54 crc kubenswrapper[4960]: I1002 07:30:54.416096 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" event={"ID":"a786a744-833c-40f9-91a7-8b5f312a626c","Type":"ContainerStarted","Data":"0d977e91a097c65607e5a6cd0465180a67ada830f3b154d97396b4cc830f8d38"} Oct 02 07:30:54 crc kubenswrapper[4960]: I1002 07:30:54.416593 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" Oct 02 07:30:54 crc kubenswrapper[4960]: I1002 07:30:54.418159 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" event={"ID":"72b0c3f8-2a66-42da-aa83-298bda8f332d","Type":"ContainerStarted","Data":"0f19a1356bca4fa3298db86e9302733b23fbdcf8f8f5cf8f9175323d6b7a8f87"} Oct 02 07:30:54 crc kubenswrapper[4960]: I1002 07:30:54.418393 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" Oct 02 07:30:54 crc kubenswrapper[4960]: I1002 07:30:54.440314 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" podStartSLOduration=3.477805913 podStartE2EDuration="39.440294056s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.978966031 +0000 UTC m=+839.010912318" lastFinishedPulling="2025-10-02 07:30:53.941454153 +0000 UTC m=+874.973400461" observedRunningTime="2025-10-02 07:30:54.433655774 +0000 UTC m=+875.465602061" watchObservedRunningTime="2025-10-02 07:30:54.440294056 +0000 UTC m=+875.472240343" Oct 02 07:30:54 crc kubenswrapper[4960]: I1002 07:30:54.454008 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" podStartSLOduration=3.560714811 podStartE2EDuration="39.453991923s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.887872211 +0000 UTC m=+838.919818488" lastFinishedPulling="2025-10-02 07:30:53.781149313 +0000 UTC m=+874.813095600" observedRunningTime="2025-10-02 07:30:54.450809014 +0000 UTC m=+875.482755301" watchObservedRunningTime="2025-10-02 07:30:54.453991923 +0000 UTC m=+875.485938210" Oct 02 07:30:56 crc kubenswrapper[4960]: I1002 07:30:56.653816 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" Oct 02 07:30:56 crc kubenswrapper[4960]: I1002 07:30:56.657535 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-lzlkk" Oct 02 07:30:56 crc kubenswrapper[4960]: I1002 07:30:56.668117 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmrs5" Oct 02 07:30:56 crc kubenswrapper[4960]: I1002 07:30:56.885089 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-fkgnk" Oct 02 07:30:56 crc kubenswrapper[4960]: I1002 07:30:56.977330 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 07:30:57 crc kubenswrapper[4960]: I1002 07:30:57.011711 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:57 crc kubenswrapper[4960]: I1002 07:30:57.018464 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-787874f5b776jcd" Oct 02 07:30:57 crc kubenswrapper[4960]: I1002 07:30:57.032427 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-69k7g" Oct 02 07:30:58 crc kubenswrapper[4960]: I1002 07:30:58.453952 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" event={"ID":"cf515440-a75b-4f32-8c08-2f108da16b56","Type":"ContainerStarted","Data":"66e42f1fce671246f2241367e65a2e1cc662d6d23de9b7e70804132104e665cd"} Oct 02 07:30:58 crc kubenswrapper[4960]: I1002 07:30:58.454594 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:30:58 crc kubenswrapper[4960]: I1002 07:30:58.476421 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" podStartSLOduration=3.957781066 podStartE2EDuration="43.476390456s" podCreationTimestamp="2025-10-02 07:30:15 +0000 UTC" firstStartedPulling="2025-10-02 07:30:17.954799474 +0000 UTC m=+838.986745761" lastFinishedPulling="2025-10-02 07:30:57.473408844 +0000 UTC m=+878.505355151" observedRunningTime="2025-10-02 07:30:58.469876157 +0000 UTC m=+879.501822444" watchObservedRunningTime="2025-10-02 07:30:58.476390456 +0000 UTC m=+879.508336773" Oct 02 07:31:03 crc kubenswrapper[4960]: I1002 07:31:03.426923 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-drs8j" Oct 02 07:31:03 crc kubenswrapper[4960]: I1002 07:31:03.492673 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-drs8j"] Oct 02 07:31:03 crc kubenswrapper[4960]: I1002 07:31:03.560443 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-549bm"] Oct 02 07:31:03 crc kubenswrapper[4960]: I1002 07:31:03.560774 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-549bm" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="registry-server" containerID="cri-o://544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267" gracePeriod=2 Oct 02 07:31:03 crc kubenswrapper[4960]: I1002 07:31:03.977894 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.045694 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-574gc\" (UniqueName: \"kubernetes.io/projected/8dd9b498-cece-42f6-88b7-65b998f421b0-kube-api-access-574gc\") pod \"8dd9b498-cece-42f6-88b7-65b998f421b0\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.047078 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-catalog-content\") pod \"8dd9b498-cece-42f6-88b7-65b998f421b0\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.047220 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-utilities\") pod \"8dd9b498-cece-42f6-88b7-65b998f421b0\" (UID: \"8dd9b498-cece-42f6-88b7-65b998f421b0\") " Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.048156 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-utilities" (OuterVolumeSpecName: "utilities") pod "8dd9b498-cece-42f6-88b7-65b998f421b0" (UID: "8dd9b498-cece-42f6-88b7-65b998f421b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.053493 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dd9b498-cece-42f6-88b7-65b998f421b0-kube-api-access-574gc" (OuterVolumeSpecName: "kube-api-access-574gc") pod "8dd9b498-cece-42f6-88b7-65b998f421b0" (UID: "8dd9b498-cece-42f6-88b7-65b998f421b0"). InnerVolumeSpecName "kube-api-access-574gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.091757 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dd9b498-cece-42f6-88b7-65b998f421b0" (UID: "8dd9b498-cece-42f6-88b7-65b998f421b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.148939 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-574gc\" (UniqueName: \"kubernetes.io/projected/8dd9b498-cece-42f6-88b7-65b998f421b0-kube-api-access-574gc\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.149004 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.149017 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd9b498-cece-42f6-88b7-65b998f421b0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.516947 4960 generic.go:334] "Generic (PLEG): container finished" podID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerID="544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267" exitCode=0 Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.517023 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-549bm" event={"ID":"8dd9b498-cece-42f6-88b7-65b998f421b0","Type":"ContainerDied","Data":"544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267"} Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.517074 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-549bm" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.518259 4960 scope.go:117] "RemoveContainer" containerID="544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.518184 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-549bm" event={"ID":"8dd9b498-cece-42f6-88b7-65b998f421b0","Type":"ContainerDied","Data":"6dcdc6bb5007d2a664eaeb5e63340b2bbedbe0970e71e3ff80e9e187b2485772"} Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.549646 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-549bm"] Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.553876 4960 scope.go:117] "RemoveContainer" containerID="a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.558286 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-549bm"] Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.573613 4960 scope.go:117] "RemoveContainer" containerID="be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.605221 4960 scope.go:117] "RemoveContainer" containerID="544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267" Oct 02 07:31:04 crc kubenswrapper[4960]: E1002 07:31:04.606953 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267\": container with ID starting with 544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267 not found: ID does not exist" containerID="544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.607080 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267"} err="failed to get container status \"544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267\": rpc error: code = NotFound desc = could not find container \"544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267\": container with ID starting with 544500c8e0b13d8f6ac721069917c8320a91be7ffb0af59a89f2a73189c04267 not found: ID does not exist" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.607159 4960 scope.go:117] "RemoveContainer" containerID="a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb" Oct 02 07:31:04 crc kubenswrapper[4960]: E1002 07:31:04.607617 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb\": container with ID starting with a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb not found: ID does not exist" containerID="a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.607710 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb"} err="failed to get container status \"a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb\": rpc error: code = NotFound desc = could not find container \"a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb\": container with ID starting with a19f3f531b6de6d7ddcde3ea4e385b4806b92a12fe96b65f46e6707bd136fefb not found: ID does not exist" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.607786 4960 scope.go:117] "RemoveContainer" containerID="be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71" Oct 02 07:31:04 crc kubenswrapper[4960]: E1002 07:31:04.608114 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71\": container with ID starting with be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71 not found: ID does not exist" containerID="be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71" Oct 02 07:31:04 crc kubenswrapper[4960]: I1002 07:31:04.608211 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71"} err="failed to get container status \"be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71\": rpc error: code = NotFound desc = could not find container \"be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71\": container with ID starting with be0fa7d93b3a0eee057e0916e3857ad044923bb23f17140fa6cf3ebe8e5c5e71 not found: ID does not exist" Oct 02 07:31:06 crc kubenswrapper[4960]: I1002 07:31:06.196344 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-cd9cz" Oct 02 07:31:06 crc kubenswrapper[4960]: I1002 07:31:06.345848 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" path="/var/lib/kubelet/pods/8dd9b498-cece-42f6-88b7-65b998f421b0/volumes" Oct 02 07:31:06 crc kubenswrapper[4960]: I1002 07:31:06.540626 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-hbzp7" Oct 02 07:31:06 crc kubenswrapper[4960]: I1002 07:31:06.602708 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-lk5kd" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.787022 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7968455899-f2fxb"] Oct 02 07:31:21 crc kubenswrapper[4960]: E1002 07:31:21.788069 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="registry-server" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788088 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="registry-server" Oct 02 07:31:21 crc kubenswrapper[4960]: E1002 07:31:21.788160 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="extract-content" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788170 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="extract-content" Oct 02 07:31:21 crc kubenswrapper[4960]: E1002 07:31:21.788203 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="extract-utilities" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788213 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="extract-utilities" Oct 02 07:31:21 crc kubenswrapper[4960]: E1002 07:31:21.788230 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="extract-utilities" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788240 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="extract-utilities" Oct 02 07:31:21 crc kubenswrapper[4960]: E1002 07:31:21.788278 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="extract-content" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788288 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="extract-content" Oct 02 07:31:21 crc kubenswrapper[4960]: E1002 07:31:21.788306 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="registry-server" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788316 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="registry-server" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788513 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dd9b498-cece-42f6-88b7-65b998f421b0" containerName="registry-server" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.788543 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4be745-c2cd-4a2b-844b-67572c49ca08" containerName="registry-server" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.789478 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.792847 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.793229 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.793539 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-zrljm" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.793707 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.801304 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7968455899-f2fxb"] Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.901685 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-574db999b5-qj4nb"] Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.903772 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.909845 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.944863 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-qj4nb"] Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.961967 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-config\") pod \"dnsmasq-dns-7968455899-f2fxb\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:21 crc kubenswrapper[4960]: I1002 07:31:21.962107 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m5xh\" (UniqueName: \"kubernetes.io/projected/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-kube-api-access-7m5xh\") pod \"dnsmasq-dns-7968455899-f2fxb\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.065430 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m5xh\" (UniqueName: \"kubernetes.io/projected/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-kube-api-access-7m5xh\") pod \"dnsmasq-dns-7968455899-f2fxb\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.065736 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-config\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.065862 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjg7c\" (UniqueName: \"kubernetes.io/projected/7a58fcf5-0861-445b-a3ca-0841753b55f3-kube-api-access-kjg7c\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.066029 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-dns-svc\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.066151 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-config\") pod \"dnsmasq-dns-7968455899-f2fxb\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.067525 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-config\") pod \"dnsmasq-dns-7968455899-f2fxb\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.087783 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m5xh\" (UniqueName: \"kubernetes.io/projected/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-kube-api-access-7m5xh\") pod \"dnsmasq-dns-7968455899-f2fxb\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.119701 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.170663 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-config\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.170815 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjg7c\" (UniqueName: \"kubernetes.io/projected/7a58fcf5-0861-445b-a3ca-0841753b55f3-kube-api-access-kjg7c\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.170948 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-dns-svc\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.183713 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-config\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.184050 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-dns-svc\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.196804 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjg7c\" (UniqueName: \"kubernetes.io/projected/7a58fcf5-0861-445b-a3ca-0841753b55f3-kube-api-access-kjg7c\") pod \"dnsmasq-dns-574db999b5-qj4nb\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.238883 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.575714 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7968455899-f2fxb"] Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.584595 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.695117 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7968455899-f2fxb" event={"ID":"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84","Type":"ContainerStarted","Data":"81c76b385355f6efa07405c49ece4c7c6dc46ca3c5b6d161d338f4454f2376ad"} Oct 02 07:31:22 crc kubenswrapper[4960]: I1002 07:31:22.700722 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-qj4nb"] Oct 02 07:31:22 crc kubenswrapper[4960]: W1002 07:31:22.703635 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a58fcf5_0861_445b_a3ca_0841753b55f3.slice/crio-97665d24e59a65e07702d2ba31c1e39afac7e4f447f710edd4e41b16efc3365e WatchSource:0}: Error finding container 97665d24e59a65e07702d2ba31c1e39afac7e4f447f710edd4e41b16efc3365e: Status 404 returned error can't find the container with id 97665d24e59a65e07702d2ba31c1e39afac7e4f447f710edd4e41b16efc3365e Oct 02 07:31:23 crc kubenswrapper[4960]: I1002 07:31:23.705129 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574db999b5-qj4nb" event={"ID":"7a58fcf5-0861-445b-a3ca-0841753b55f3","Type":"ContainerStarted","Data":"97665d24e59a65e07702d2ba31c1e39afac7e4f447f710edd4e41b16efc3365e"} Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.675370 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-qj4nb"] Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.723337 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-scccg"] Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.725036 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.729229 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-scccg"] Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.818615 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-config\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.818666 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmqmd\" (UniqueName: \"kubernetes.io/projected/b847ef95-9ec1-44ae-b395-f1acd7bbea62-kube-api-access-cmqmd\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.818694 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-dns-svc\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.921071 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-config\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.921134 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmqmd\" (UniqueName: \"kubernetes.io/projected/b847ef95-9ec1-44ae-b395-f1acd7bbea62-kube-api-access-cmqmd\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.921157 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-dns-svc\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.922344 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-dns-svc\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.922612 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-config\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.953113 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmqmd\" (UniqueName: \"kubernetes.io/projected/b847ef95-9ec1-44ae-b395-f1acd7bbea62-kube-api-access-cmqmd\") pod \"dnsmasq-dns-7878bd5965-scccg\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:24 crc kubenswrapper[4960]: I1002 07:31:24.996853 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7968455899-f2fxb"] Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.027861 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-qnq6h"] Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.029932 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.038065 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-qnq6h"] Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.056923 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.130685 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64tfm\" (UniqueName: \"kubernetes.io/projected/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-kube-api-access-64tfm\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.130781 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-config\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.130814 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-dns-svc\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.232389 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-dns-svc\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.232515 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64tfm\" (UniqueName: \"kubernetes.io/projected/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-kube-api-access-64tfm\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.232552 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-config\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.234198 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-dns-svc\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.234430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-config\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.263700 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64tfm\" (UniqueName: \"kubernetes.io/projected/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-kube-api-access-64tfm\") pod \"dnsmasq-dns-6b5fff88f5-qnq6h\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.355921 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.412155 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-scccg"] Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.745541 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878bd5965-scccg" event={"ID":"b847ef95-9ec1-44ae-b395-f1acd7bbea62","Type":"ContainerStarted","Data":"63c6a7f9ff25adb8d6f419ed35992d63cee3ffac3f4b486329287952a3ad8ad8"} Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.868948 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.870390 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.879449 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.881551 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-s6vnf" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.880770 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.881926 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.882075 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.881432 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.881504 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.889229 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-qnq6h"] Oct 02 07:31:25 crc kubenswrapper[4960]: I1002 07:31:25.897525 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044299 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044356 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044408 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlzrr\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-kube-api-access-zlzrr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044433 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044466 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044693 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044798 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044850 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044884 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.044932 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.045157 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.133481 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.135415 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.137769 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.138292 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.142314 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.142605 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.142798 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.143021 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.143231 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7wvn9" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.146999 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147064 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147097 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147249 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlzrr\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-kube-api-access-zlzrr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147373 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147402 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147463 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147485 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147525 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147549 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147567 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.147695 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.148714 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.149359 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.149822 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.150259 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.151482 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.152834 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.161664 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.166368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.167199 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.168529 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlzrr\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-kube-api-access-zlzrr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.176177 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.180719 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.204726 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249490 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249588 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249627 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249669 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36c68533-435a-4bb7-b338-4cd484b155f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249693 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249804 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249860 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249883 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36c68533-435a-4bb7-b338-4cd484b155f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.249984 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc5zm\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-kube-api-access-xc5zm\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.250055 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.250075 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-config-data\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352232 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352319 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352349 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352373 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36c68533-435a-4bb7-b338-4cd484b155f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352393 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352453 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352485 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352505 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36c68533-435a-4bb7-b338-4cd484b155f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352524 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc5zm\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-kube-api-access-xc5zm\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352551 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352569 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-config-data\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.352935 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.353298 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.354204 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.354289 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.354612 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.358112 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-config-data\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.358954 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.371731 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.372184 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36c68533-435a-4bb7-b338-4cd484b155f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.379442 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36c68533-435a-4bb7-b338-4cd484b155f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.390298 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc5zm\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-kube-api-access-xc5zm\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.401306 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.536497 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:31:26 crc kubenswrapper[4960]: I1002 07:31:26.758131 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" event={"ID":"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e","Type":"ContainerStarted","Data":"7fad62d309c49af1558e4a198083b680a766d4650571dcb0ca5c94401fe61ae3"} Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.459012 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.460719 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.462781 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.464747 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.464820 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xz7pb" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.470183 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.472737 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.476485 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.480106 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.598729 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-kolla-config\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.598812 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.598895 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-config-data-default\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.598947 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-secrets\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.599114 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.599712 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.599790 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.599942 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.601151 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crxvx\" (UniqueName: \"kubernetes.io/projected/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-kube-api-access-crxvx\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704045 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crxvx\" (UniqueName: \"kubernetes.io/projected/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-kube-api-access-crxvx\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704571 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-kolla-config\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704631 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704669 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-config-data-default\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704701 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-secrets\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704747 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704822 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704881 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.704936 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.705577 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.705720 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-kolla-config\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.706508 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.707055 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-config-data-default\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.708246 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.712559 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.713410 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-secrets\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.723146 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.726653 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crxvx\" (UniqueName: \"kubernetes.io/projected/3be247d9-d9ed-49f1-9013-2ad50b6e4ed8-kube-api-access-crxvx\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.732584 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8\") " pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.789446 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.861495 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.864165 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.866373 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.867195 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hs7mn" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.868013 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.868140 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 07:31:28 crc kubenswrapper[4960]: I1002 07:31:28.874405 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.010670 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.010738 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.010819 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ea27c7c-a834-41de-9a07-3f0611c3bc23-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.010881 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.010940 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.010999 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.011048 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x69pb\" (UniqueName: \"kubernetes.io/projected/1ea27c7c-a834-41de-9a07-3f0611c3bc23-kube-api-access-x69pb\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.011086 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.011110 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113400 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113486 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113547 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x69pb\" (UniqueName: \"kubernetes.io/projected/1ea27c7c-a834-41de-9a07-3f0611c3bc23-kube-api-access-x69pb\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113573 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113595 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113661 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113692 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113719 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ea27c7c-a834-41de-9a07-3f0611c3bc23-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.113743 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.114689 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.115038 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.115205 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ea27c7c-a834-41de-9a07-3f0611c3bc23-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.115278 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.116113 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea27c7c-a834-41de-9a07-3f0611c3bc23-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.122669 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.123143 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.124280 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1ea27c7c-a834-41de-9a07-3f0611c3bc23-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.137106 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x69pb\" (UniqueName: \"kubernetes.io/projected/1ea27c7c-a834-41de-9a07-3f0611c3bc23-kube-api-access-x69pb\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.148121 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ea27c7c-a834-41de-9a07-3f0611c3bc23\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.149735 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.149810 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.186904 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.532509 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.533633 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.535799 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.536483 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.537014 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-n6ff2" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.562769 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.622210 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fded14c9-2fd8-46e6-83a9-84c4ca056120-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.622262 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fded14c9-2fd8-46e6-83a9-84c4ca056120-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.622527 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v264p\" (UniqueName: \"kubernetes.io/projected/fded14c9-2fd8-46e6-83a9-84c4ca056120-kube-api-access-v264p\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.622793 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fded14c9-2fd8-46e6-83a9-84c4ca056120-kolla-config\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.622901 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fded14c9-2fd8-46e6-83a9-84c4ca056120-config-data\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.724791 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v264p\" (UniqueName: \"kubernetes.io/projected/fded14c9-2fd8-46e6-83a9-84c4ca056120-kube-api-access-v264p\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.724876 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fded14c9-2fd8-46e6-83a9-84c4ca056120-kolla-config\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.724907 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fded14c9-2fd8-46e6-83a9-84c4ca056120-config-data\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.724953 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fded14c9-2fd8-46e6-83a9-84c4ca056120-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.725059 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fded14c9-2fd8-46e6-83a9-84c4ca056120-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.725944 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fded14c9-2fd8-46e6-83a9-84c4ca056120-kolla-config\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.726128 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fded14c9-2fd8-46e6-83a9-84c4ca056120-config-data\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.731760 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fded14c9-2fd8-46e6-83a9-84c4ca056120-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.735631 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fded14c9-2fd8-46e6-83a9-84c4ca056120-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.743172 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v264p\" (UniqueName: \"kubernetes.io/projected/fded14c9-2fd8-46e6-83a9-84c4ca056120-kube-api-access-v264p\") pod \"memcached-0\" (UID: \"fded14c9-2fd8-46e6-83a9-84c4ca056120\") " pod="openstack/memcached-0" Oct 02 07:31:29 crc kubenswrapper[4960]: I1002 07:31:29.849902 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.264606 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.265989 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.268584 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-dbnxj" Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.295780 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.359751 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plqxz\" (UniqueName: \"kubernetes.io/projected/7f55241f-c680-4282-9d60-2724a8ba9e7a-kube-api-access-plqxz\") pod \"kube-state-metrics-0\" (UID: \"7f55241f-c680-4282-9d60-2724a8ba9e7a\") " pod="openstack/kube-state-metrics-0" Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.463529 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plqxz\" (UniqueName: \"kubernetes.io/projected/7f55241f-c680-4282-9d60-2724a8ba9e7a-kube-api-access-plqxz\") pod \"kube-state-metrics-0\" (UID: \"7f55241f-c680-4282-9d60-2724a8ba9e7a\") " pod="openstack/kube-state-metrics-0" Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.488394 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plqxz\" (UniqueName: \"kubernetes.io/projected/7f55241f-c680-4282-9d60-2724a8ba9e7a-kube-api-access-plqxz\") pod \"kube-state-metrics-0\" (UID: \"7f55241f-c680-4282-9d60-2724a8ba9e7a\") " pod="openstack/kube-state-metrics-0" Oct 02 07:31:31 crc kubenswrapper[4960]: I1002 07:31:31.588037 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:31:33 crc kubenswrapper[4960]: I1002 07:31:33.470693 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.256461 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.258174 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.265908 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.266802 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.266869 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.266900 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-dkcvw" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.267209 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.274535 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353286 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/225801bc-ca07-4fa8-8d5c-1d56cfef957b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353356 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353390 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j6vp\" (UniqueName: \"kubernetes.io/projected/225801bc-ca07-4fa8-8d5c-1d56cfef957b-kube-api-access-6j6vp\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353486 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/225801bc-ca07-4fa8-8d5c-1d56cfef957b-config\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353550 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353581 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/225801bc-ca07-4fa8-8d5c-1d56cfef957b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.353609 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455160 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/225801bc-ca07-4fa8-8d5c-1d56cfef957b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455248 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455299 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/225801bc-ca07-4fa8-8d5c-1d56cfef957b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455317 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455339 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j6vp\" (UniqueName: \"kubernetes.io/projected/225801bc-ca07-4fa8-8d5c-1d56cfef957b-kube-api-access-6j6vp\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455468 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/225801bc-ca07-4fa8-8d5c-1d56cfef957b-config\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455497 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.455947 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.457819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/225801bc-ca07-4fa8-8d5c-1d56cfef957b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.458136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/225801bc-ca07-4fa8-8d5c-1d56cfef957b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.461658 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/225801bc-ca07-4fa8-8d5c-1d56cfef957b-config\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.473284 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.479915 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.482804 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/225801bc-ca07-4fa8-8d5c-1d56cfef957b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.492888 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j6vp\" (UniqueName: \"kubernetes.io/projected/225801bc-ca07-4fa8-8d5c-1d56cfef957b-kube-api-access-6j6vp\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.552498 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-krnzn"] Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.553821 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.559172 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"225801bc-ca07-4fa8-8d5c-1d56cfef957b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.564364 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.564474 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-hshz6" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.564633 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.582801 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krnzn"] Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.598874 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.663987 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-run\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.664074 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qskk\" (UniqueName: \"kubernetes.io/projected/ad314560-608d-471a-a6f6-659b9c0755d6-kube-api-access-5qskk\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.664108 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-run-ovn\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.664124 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-log-ovn\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.664153 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad314560-608d-471a-a6f6-659b9c0755d6-combined-ca-bundle\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.664173 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad314560-608d-471a-a6f6-659b9c0755d6-ovn-controller-tls-certs\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.664413 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad314560-608d-471a-a6f6-659b9c0755d6-scripts\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.698912 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-f9tj4"] Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.703278 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.727608 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-f9tj4"] Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qskk\" (UniqueName: \"kubernetes.io/projected/ad314560-608d-471a-a6f6-659b9c0755d6-kube-api-access-5qskk\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766276 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-run-ovn\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766302 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-log\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766319 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-log-ovn\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766349 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad314560-608d-471a-a6f6-659b9c0755d6-combined-ca-bundle\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766369 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad314560-608d-471a-a6f6-659b9c0755d6-ovn-controller-tls-certs\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766385 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-scripts\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766414 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad314560-608d-471a-a6f6-659b9c0755d6-scripts\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766452 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfc5m\" (UniqueName: \"kubernetes.io/projected/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-kube-api-access-sfc5m\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766479 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-run\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766498 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-etc-ovs\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766516 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-run\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.766536 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-lib\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.767358 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-run-ovn\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.767487 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-log-ovn\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.769105 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ad314560-608d-471a-a6f6-659b9c0755d6-var-run\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.772407 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad314560-608d-471a-a6f6-659b9c0755d6-combined-ca-bundle\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.772877 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad314560-608d-471a-a6f6-659b9c0755d6-ovn-controller-tls-certs\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.773808 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad314560-608d-471a-a6f6-659b9c0755d6-scripts\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.796789 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qskk\" (UniqueName: \"kubernetes.io/projected/ad314560-608d-471a-a6f6-659b9c0755d6-kube-api-access-5qskk\") pod \"ovn-controller-krnzn\" (UID: \"ad314560-608d-471a-a6f6-659b9c0755d6\") " pod="openstack/ovn-controller-krnzn" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.868487 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-lib\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.868587 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-log\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.868627 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-scripts\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.868680 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfc5m\" (UniqueName: \"kubernetes.io/projected/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-kube-api-access-sfc5m\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.868719 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-run\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.868742 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-etc-ovs\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.869083 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-etc-ovs\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.869254 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-lib\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.869356 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-log\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.871426 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-scripts\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.873476 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-var-run\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.891669 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfc5m\" (UniqueName: \"kubernetes.io/projected/bff7cbe3-e11d-478e-8ce8-1a03becd8ac8-kube-api-access-sfc5m\") pod \"ovn-controller-ovs-f9tj4\" (UID: \"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8\") " pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:35 crc kubenswrapper[4960]: I1002 07:31:35.935865 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn" Oct 02 07:31:36 crc kubenswrapper[4960]: I1002 07:31:36.027749 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:38 crc kubenswrapper[4960]: W1002 07:31:38.574391 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ea27c7c_a834_41de_9a07_3f0611c3bc23.slice/crio-8cba076f3d04d55fbd23c277ed1f8e2d1f3a5a0c142da176dd554cef5c91e607 WatchSource:0}: Error finding container 8cba076f3d04d55fbd23c277ed1f8e2d1f3a5a0c142da176dd554cef5c91e607: Status 404 returned error can't find the container with id 8cba076f3d04d55fbd23c277ed1f8e2d1f3a5a0c142da176dd554cef5c91e607 Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.871791 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ea27c7c-a834-41de-9a07-3f0611c3bc23","Type":"ContainerStarted","Data":"8cba076f3d04d55fbd23c277ed1f8e2d1f3a5a0c142da176dd554cef5c91e607"} Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.952309 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.956323 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.959465 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qb4tw" Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.959812 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.960102 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.960917 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 07:31:38 crc kubenswrapper[4960]: I1002 07:31:38.962214 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.023536 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.032756 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.032806 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.032851 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jnn4\" (UniqueName: \"kubernetes.io/projected/67a98bd9-8fba-44c5-9229-de8a1053d9d6-kube-api-access-6jnn4\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.032877 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.032904 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.032924 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67a98bd9-8fba-44c5-9229-de8a1053d9d6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.032961 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67a98bd9-8fba-44c5-9229-de8a1053d9d6-config\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.033015 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67a98bd9-8fba-44c5-9229-de8a1053d9d6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136307 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136364 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136413 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jnn4\" (UniqueName: \"kubernetes.io/projected/67a98bd9-8fba-44c5-9229-de8a1053d9d6-kube-api-access-6jnn4\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136442 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136476 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136505 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67a98bd9-8fba-44c5-9229-de8a1053d9d6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136548 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67a98bd9-8fba-44c5-9229-de8a1053d9d6-config\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.136596 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67a98bd9-8fba-44c5-9229-de8a1053d9d6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.137323 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67a98bd9-8fba-44c5-9229-de8a1053d9d6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.137458 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.138361 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67a98bd9-8fba-44c5-9229-de8a1053d9d6-config\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.138521 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67a98bd9-8fba-44c5-9229-de8a1053d9d6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.146547 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.148265 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.149760 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67a98bd9-8fba-44c5-9229-de8a1053d9d6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.156829 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jnn4\" (UniqueName: \"kubernetes.io/projected/67a98bd9-8fba-44c5-9229-de8a1053d9d6-kube-api-access-6jnn4\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.169988 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"67a98bd9-8fba-44c5-9229-de8a1053d9d6\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.310585 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.604194 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.604338 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7m5xh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7968455899-f2fxb_openstack(7f8dc5d4-00c3-44ad-a666-c3e5560c1a84): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.604674 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.604768 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kjg7c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-574db999b5-qj4nb_openstack(7a58fcf5-0861-445b-a3ca-0841753b55f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.606165 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-574db999b5-qj4nb" podUID="7a58fcf5-0861-445b-a3ca-0841753b55f3" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.606358 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7968455899-f2fxb" podUID="7f8dc5d4-00c3-44ad-a666-c3e5560c1a84" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.618208 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.618503 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cmqmd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7878bd5965-scccg_openstack(b847ef95-9ec1-44ae-b395-f1acd7bbea62): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:31:39 crc kubenswrapper[4960]: E1002 07:31:39.620067 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7878bd5965-scccg" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" Oct 02 07:31:39 crc kubenswrapper[4960]: I1002 07:31:39.881962 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fded14c9-2fd8-46e6-83a9-84c4ca056120","Type":"ContainerStarted","Data":"a9aecf04f8499a2cfe067d7a142dd442069a7a19202809755228ec5ab7799428"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.064284 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:31:40 crc kubenswrapper[4960]: W1002 07:31:40.086548 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8c7a2f9_57bf_4520_b078_ce75a2e051ad.slice/crio-979d1f7ec5b1271b4a68630ac106b23942e7e072c3ce1ed0ffc50da9990dab23 WatchSource:0}: Error finding container 979d1f7ec5b1271b4a68630ac106b23942e7e072c3ce1ed0ffc50da9990dab23: Status 404 returned error can't find the container with id 979d1f7ec5b1271b4a68630ac106b23942e7e072c3ce1ed0ffc50da9990dab23 Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.129544 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 07:31:40 crc kubenswrapper[4960]: W1002 07:31:40.211991 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3be247d9_d9ed_49f1_9013_2ad50b6e4ed8.slice/crio-f31e4ce9a44a21c814058c3425dffcf1b9e54bdda02451ff1258107d3bad3bc0 WatchSource:0}: Error finding container f31e4ce9a44a21c814058c3425dffcf1b9e54bdda02451ff1258107d3bad3bc0: Status 404 returned error can't find the container with id f31e4ce9a44a21c814058c3425dffcf1b9e54bdda02451ff1258107d3bad3bc0 Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.477517 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.525099 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.529525 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.530433 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:31:40 crc kubenswrapper[4960]: W1002 07:31:40.549244 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f55241f_c680_4282_9d60_2724a8ba9e7a.slice/crio-a9284cfe1b204fc871c0e83b914dd0e12e54a7730b1450f9b7c3cdd83cbebfd9 WatchSource:0}: Error finding container a9284cfe1b204fc871c0e83b914dd0e12e54a7730b1450f9b7c3cdd83cbebfd9: Status 404 returned error can't find the container with id a9284cfe1b204fc871c0e83b914dd0e12e54a7730b1450f9b7c3cdd83cbebfd9 Oct 02 07:31:40 crc kubenswrapper[4960]: W1002 07:31:40.560882 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36c68533_435a_4bb7_b338_4cd484b155f6.slice/crio-a1fffe3454e9faced5855e44fbec91b55d8e3a9481801baa93f7bac98a847031 WatchSource:0}: Error finding container a1fffe3454e9faced5855e44fbec91b55d8e3a9481801baa93f7bac98a847031: Status 404 returned error can't find the container with id a1fffe3454e9faced5855e44fbec91b55d8e3a9481801baa93f7bac98a847031 Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.609257 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-config\") pod \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.609398 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m5xh\" (UniqueName: \"kubernetes.io/projected/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-kube-api-access-7m5xh\") pod \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\" (UID: \"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84\") " Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.609571 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-config\") pod \"7a58fcf5-0861-445b-a3ca-0841753b55f3\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.609645 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-dns-svc\") pod \"7a58fcf5-0861-445b-a3ca-0841753b55f3\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.609705 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjg7c\" (UniqueName: \"kubernetes.io/projected/7a58fcf5-0861-445b-a3ca-0841753b55f3-kube-api-access-kjg7c\") pod \"7a58fcf5-0861-445b-a3ca-0841753b55f3\" (UID: \"7a58fcf5-0861-445b-a3ca-0841753b55f3\") " Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.609959 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-config" (OuterVolumeSpecName: "config") pod "7f8dc5d4-00c3-44ad-a666-c3e5560c1a84" (UID: "7f8dc5d4-00c3-44ad-a666-c3e5560c1a84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.610218 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.610458 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-config" (OuterVolumeSpecName: "config") pod "7a58fcf5-0861-445b-a3ca-0841753b55f3" (UID: "7a58fcf5-0861-445b-a3ca-0841753b55f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.611613 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a58fcf5-0861-445b-a3ca-0841753b55f3" (UID: "7a58fcf5-0861-445b-a3ca-0841753b55f3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.618931 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a58fcf5-0861-445b-a3ca-0841753b55f3-kube-api-access-kjg7c" (OuterVolumeSpecName: "kube-api-access-kjg7c") pod "7a58fcf5-0861-445b-a3ca-0841753b55f3" (UID: "7a58fcf5-0861-445b-a3ca-0841753b55f3"). InnerVolumeSpecName "kube-api-access-kjg7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.619090 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-kube-api-access-7m5xh" (OuterVolumeSpecName: "kube-api-access-7m5xh") pod "7f8dc5d4-00c3-44ad-a666-c3e5560c1a84" (UID: "7f8dc5d4-00c3-44ad-a666-c3e5560c1a84"). InnerVolumeSpecName "kube-api-access-7m5xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.646958 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krnzn"] Oct 02 07:31:40 crc kubenswrapper[4960]: W1002 07:31:40.668754 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad314560_608d_471a_a6f6_659b9c0755d6.slice/crio-715491f5e22fa4b9bf209981c331028c5700eab306da107e1e951117cdb2303a WatchSource:0}: Error finding container 715491f5e22fa4b9bf209981c331028c5700eab306da107e1e951117cdb2303a: Status 404 returned error can't find the container with id 715491f5e22fa4b9bf209981c331028c5700eab306da107e1e951117cdb2303a Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.712405 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m5xh\" (UniqueName: \"kubernetes.io/projected/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84-kube-api-access-7m5xh\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.712876 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.712890 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a58fcf5-0861-445b-a3ca-0841753b55f3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.712903 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjg7c\" (UniqueName: \"kubernetes.io/projected/7a58fcf5-0861-445b-a3ca-0841753b55f3-kube-api-access-kjg7c\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.752184 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.821386 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 07:31:40 crc kubenswrapper[4960]: W1002 07:31:40.851554 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67a98bd9_8fba_44c5_9229_de8a1053d9d6.slice/crio-3661e2d2544fb746f8eb34d1f177b487c4b49c99bb1568fba8e03de9370198cc WatchSource:0}: Error finding container 3661e2d2544fb746f8eb34d1f177b487c4b49c99bb1568fba8e03de9370198cc: Status 404 returned error can't find the container with id 3661e2d2544fb746f8eb34d1f177b487c4b49c99bb1568fba8e03de9370198cc Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.890420 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn" event={"ID":"ad314560-608d-471a-a6f6-659b9c0755d6","Type":"ContainerStarted","Data":"715491f5e22fa4b9bf209981c331028c5700eab306da107e1e951117cdb2303a"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.892137 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"67a98bd9-8fba-44c5-9229-de8a1053d9d6","Type":"ContainerStarted","Data":"3661e2d2544fb746f8eb34d1f177b487c4b49c99bb1568fba8e03de9370198cc"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.893797 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574db999b5-qj4nb" event={"ID":"7a58fcf5-0861-445b-a3ca-0841753b55f3","Type":"ContainerDied","Data":"97665d24e59a65e07702d2ba31c1e39afac7e4f447f710edd4e41b16efc3365e"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.893804 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-qj4nb" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.895786 4960 generic.go:334] "Generic (PLEG): container finished" podID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerID="4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6" exitCode=0 Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.895842 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" event={"ID":"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e","Type":"ContainerDied","Data":"4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.897823 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"225801bc-ca07-4fa8-8d5c-1d56cfef957b","Type":"ContainerStarted","Data":"399e3c50a82bb0cb50cecda2afcaf43d9b8cc4f977bb09d7989e8cf13234a9a5"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.900253 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8","Type":"ContainerStarted","Data":"f31e4ce9a44a21c814058c3425dffcf1b9e54bdda02451ff1258107d3bad3bc0"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.901635 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8c7a2f9-57bf-4520-b078-ce75a2e051ad","Type":"ContainerStarted","Data":"979d1f7ec5b1271b4a68630ac106b23942e7e072c3ce1ed0ffc50da9990dab23"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.904126 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"36c68533-435a-4bb7-b338-4cd484b155f6","Type":"ContainerStarted","Data":"a1fffe3454e9faced5855e44fbec91b55d8e3a9481801baa93f7bac98a847031"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.906263 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7f55241f-c680-4282-9d60-2724a8ba9e7a","Type":"ContainerStarted","Data":"a9284cfe1b204fc871c0e83b914dd0e12e54a7730b1450f9b7c3cdd83cbebfd9"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.911684 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7968455899-f2fxb" event={"ID":"7f8dc5d4-00c3-44ad-a666-c3e5560c1a84","Type":"ContainerDied","Data":"81c76b385355f6efa07405c49ece4c7c6dc46ca3c5b6d161d338f4454f2376ad"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.911714 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-f2fxb" Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.913962 4960 generic.go:334] "Generic (PLEG): container finished" podID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerID="bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8" exitCode=0 Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.914053 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878bd5965-scccg" event={"ID":"b847ef95-9ec1-44ae-b395-f1acd7bbea62","Type":"ContainerDied","Data":"bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8"} Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.982274 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-qj4nb"] Oct 02 07:31:40 crc kubenswrapper[4960]: I1002 07:31:40.996222 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-qj4nb"] Oct 02 07:31:41 crc kubenswrapper[4960]: I1002 07:31:41.012289 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7968455899-f2fxb"] Oct 02 07:31:41 crc kubenswrapper[4960]: I1002 07:31:41.017615 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7968455899-f2fxb"] Oct 02 07:31:41 crc kubenswrapper[4960]: I1002 07:31:41.275655 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-f9tj4"] Oct 02 07:31:41 crc kubenswrapper[4960]: I1002 07:31:41.923750 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f9tj4" event={"ID":"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8","Type":"ContainerStarted","Data":"5ee99c771d0acf55ff8d533919de8e7b84e931c3d661fb5b3c818bac3fd51d8a"} Oct 02 07:31:42 crc kubenswrapper[4960]: I1002 07:31:42.339255 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a58fcf5-0861-445b-a3ca-0841753b55f3" path="/var/lib/kubelet/pods/7a58fcf5-0861-445b-a3ca-0841753b55f3/volumes" Oct 02 07:31:42 crc kubenswrapper[4960]: I1002 07:31:42.339729 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f8dc5d4-00c3-44ad-a666-c3e5560c1a84" path="/var/lib/kubelet/pods/7f8dc5d4-00c3-44ad-a666-c3e5560c1a84/volumes" Oct 02 07:31:43 crc kubenswrapper[4960]: I1002 07:31:43.945184 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878bd5965-scccg" event={"ID":"b847ef95-9ec1-44ae-b395-f1acd7bbea62","Type":"ContainerStarted","Data":"cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32"} Oct 02 07:31:43 crc kubenswrapper[4960]: I1002 07:31:43.945785 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:43 crc kubenswrapper[4960]: I1002 07:31:43.950631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" event={"ID":"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e","Type":"ContainerStarted","Data":"67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a"} Oct 02 07:31:43 crc kubenswrapper[4960]: I1002 07:31:43.950784 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:43 crc kubenswrapper[4960]: I1002 07:31:43.970198 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7878bd5965-scccg" podStartSLOduration=-9223372016.8846 podStartE2EDuration="19.970175245s" podCreationTimestamp="2025-10-02 07:31:24 +0000 UTC" firstStartedPulling="2025-10-02 07:31:25.446581498 +0000 UTC m=+906.478527775" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:31:43.962136815 +0000 UTC m=+924.994083112" watchObservedRunningTime="2025-10-02 07:31:43.970175245 +0000 UTC m=+925.002121532" Oct 02 07:31:43 crc kubenswrapper[4960]: I1002 07:31:43.980426 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" podStartSLOduration=5.133392208 podStartE2EDuration="18.980409044s" podCreationTimestamp="2025-10-02 07:31:25 +0000 UTC" firstStartedPulling="2025-10-02 07:31:25.911800764 +0000 UTC m=+906.943747041" lastFinishedPulling="2025-10-02 07:31:39.75881759 +0000 UTC m=+920.790763877" observedRunningTime="2025-10-02 07:31:43.97952414 +0000 UTC m=+925.011470447" watchObservedRunningTime="2025-10-02 07:31:43.980409044 +0000 UTC m=+925.012355331" Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.033553 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fded14c9-2fd8-46e6-83a9-84c4ca056120","Type":"ContainerStarted","Data":"6092ac0756ce90b8fac7c4fb80939395fede4a8d3c278155131c4633d65fdfdf"} Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.034607 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.039436 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ea27c7c-a834-41de-9a07-3f0611c3bc23","Type":"ContainerStarted","Data":"6820a6b1c7a330b8390fdc9ac46b2b9bd302606023ac81e78679364de6890e80"} Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.041854 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8","Type":"ContainerStarted","Data":"57ea5128660cbd00ec1430e2a3f0cd1755991e3ffa00e5722806d0ba687709c4"} Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.060255 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.100746 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.402713775 podStartE2EDuration="21.100721696s" podCreationTimestamp="2025-10-02 07:31:29 +0000 UTC" firstStartedPulling="2025-10-02 07:31:39.548588536 +0000 UTC m=+920.580534843" lastFinishedPulling="2025-10-02 07:31:48.246596437 +0000 UTC m=+929.278542764" observedRunningTime="2025-10-02 07:31:50.061717791 +0000 UTC m=+931.093664088" watchObservedRunningTime="2025-10-02 07:31:50.100721696 +0000 UTC m=+931.132667983" Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.357118 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:31:50 crc kubenswrapper[4960]: I1002 07:31:50.536288 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-scccg"] Oct 02 07:31:51 crc kubenswrapper[4960]: I1002 07:31:51.060438 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f9tj4" event={"ID":"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8","Type":"ContainerStarted","Data":"c5d71d7955f802bb00adaad4926e5f874e2b7017f6f802d163946e744ec5bf63"} Oct 02 07:31:51 crc kubenswrapper[4960]: I1002 07:31:51.064542 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"67a98bd9-8fba-44c5-9229-de8a1053d9d6","Type":"ContainerStarted","Data":"cfde27ab66e0f1d1372eb5d4c4c2caced6cdda8fd1982830a20bb38b41c0a43f"} Oct 02 07:31:51 crc kubenswrapper[4960]: I1002 07:31:51.073439 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"225801bc-ca07-4fa8-8d5c-1d56cfef957b","Type":"ContainerStarted","Data":"5addd866aae2468d4cd9f54e17f1d4c1514870bd4e965436e9507177eb02dd06"} Oct 02 07:31:51 crc kubenswrapper[4960]: I1002 07:31:51.076517 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7878bd5965-scccg" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerName="dnsmasq-dns" containerID="cri-o://cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32" gracePeriod=10 Oct 02 07:31:51 crc kubenswrapper[4960]: I1002 07:31:51.076926 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8c7a2f9-57bf-4520-b078-ce75a2e051ad","Type":"ContainerStarted","Data":"08856e4cd8439c757573e71be33fb3eca2f53d6f1f0a9f9e755f234506844f0f"} Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.035331 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.093632 4960 generic.go:334] "Generic (PLEG): container finished" podID="bff7cbe3-e11d-478e-8ce8-1a03becd8ac8" containerID="c5d71d7955f802bb00adaad4926e5f874e2b7017f6f802d163946e744ec5bf63" exitCode=0 Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.094682 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f9tj4" event={"ID":"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8","Type":"ContainerDied","Data":"c5d71d7955f802bb00adaad4926e5f874e2b7017f6f802d163946e744ec5bf63"} Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.110267 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"36c68533-435a-4bb7-b338-4cd484b155f6","Type":"ContainerStarted","Data":"9bd820698bae00820e1521389d7e85c9f91579903f903da17b0bbb829e24c442"} Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.119562 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7f55241f-c680-4282-9d60-2724a8ba9e7a","Type":"ContainerStarted","Data":"bce0ed98537129efbbd9480f0d9f3cdbd4ddec996d67c43034737ea6ff429c41"} Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.119968 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.125365 4960 generic.go:334] "Generic (PLEG): container finished" podID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerID="cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32" exitCode=0 Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.125427 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878bd5965-scccg" event={"ID":"b847ef95-9ec1-44ae-b395-f1acd7bbea62","Type":"ContainerDied","Data":"cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32"} Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.125448 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-scccg" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.125521 4960 scope.go:117] "RemoveContainer" containerID="cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.125465 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878bd5965-scccg" event={"ID":"b847ef95-9ec1-44ae-b395-f1acd7bbea62","Type":"ContainerDied","Data":"63c6a7f9ff25adb8d6f419ed35992d63cee3ffac3f4b486329287952a3ad8ad8"} Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.140472 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-dns-svc\") pod \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.140552 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmqmd\" (UniqueName: \"kubernetes.io/projected/b847ef95-9ec1-44ae-b395-f1acd7bbea62-kube-api-access-cmqmd\") pod \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.140635 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-config\") pod \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\" (UID: \"b847ef95-9ec1-44ae-b395-f1acd7bbea62\") " Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.155882 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b847ef95-9ec1-44ae-b395-f1acd7bbea62-kube-api-access-cmqmd" (OuterVolumeSpecName: "kube-api-access-cmqmd") pod "b847ef95-9ec1-44ae-b395-f1acd7bbea62" (UID: "b847ef95-9ec1-44ae-b395-f1acd7bbea62"). InnerVolumeSpecName "kube-api-access-cmqmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.191348 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.08485376 podStartE2EDuration="21.191321767s" podCreationTimestamp="2025-10-02 07:31:31 +0000 UTC" firstStartedPulling="2025-10-02 07:31:40.554475819 +0000 UTC m=+921.586422126" lastFinishedPulling="2025-10-02 07:31:51.660943846 +0000 UTC m=+932.692890133" observedRunningTime="2025-10-02 07:31:52.147182511 +0000 UTC m=+933.179128788" watchObservedRunningTime="2025-10-02 07:31:52.191321767 +0000 UTC m=+933.223268054" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.192425 4960 scope.go:117] "RemoveContainer" containerID="bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.231133 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-config" (OuterVolumeSpecName: "config") pod "b847ef95-9ec1-44ae-b395-f1acd7bbea62" (UID: "b847ef95-9ec1-44ae-b395-f1acd7bbea62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.231357 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b847ef95-9ec1-44ae-b395-f1acd7bbea62" (UID: "b847ef95-9ec1-44ae-b395-f1acd7bbea62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.232005 4960 scope.go:117] "RemoveContainer" containerID="cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32" Oct 02 07:31:52 crc kubenswrapper[4960]: E1002 07:31:52.233108 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32\": container with ID starting with cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32 not found: ID does not exist" containerID="cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.233165 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32"} err="failed to get container status \"cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32\": rpc error: code = NotFound desc = could not find container \"cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32\": container with ID starting with cd9adf86c23dd904a5290a1759eab1fd5a7a2cd8aac7f81cab5b123170481b32 not found: ID does not exist" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.233192 4960 scope.go:117] "RemoveContainer" containerID="bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8" Oct 02 07:31:52 crc kubenswrapper[4960]: E1002 07:31:52.234261 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8\": container with ID starting with bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8 not found: ID does not exist" containerID="bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.234312 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8"} err="failed to get container status \"bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8\": rpc error: code = NotFound desc = could not find container \"bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8\": container with ID starting with bd195bde113ae83ea6c8d12bbe6148d5a6a244c15e6370c1570241370191c1c8 not found: ID does not exist" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.243813 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.243881 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmqmd\" (UniqueName: \"kubernetes.io/projected/b847ef95-9ec1-44ae-b395-f1acd7bbea62-kube-api-access-cmqmd\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.243905 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b847ef95-9ec1-44ae-b395-f1acd7bbea62-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.448074 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-scccg"] Oct 02 07:31:52 crc kubenswrapper[4960]: I1002 07:31:52.456094 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-scccg"] Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.143510 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn" event={"ID":"ad314560-608d-471a-a6f6-659b9c0755d6","Type":"ContainerStarted","Data":"ee00eb94770586344d63ca01ec10801d0bab844b19a761746c3e6849149c1cff"} Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.144349 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-krnzn" Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.151715 4960 generic.go:334] "Generic (PLEG): container finished" podID="1ea27c7c-a834-41de-9a07-3f0611c3bc23" containerID="6820a6b1c7a330b8390fdc9ac46b2b9bd302606023ac81e78679364de6890e80" exitCode=0 Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.151865 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ea27c7c-a834-41de-9a07-3f0611c3bc23","Type":"ContainerDied","Data":"6820a6b1c7a330b8390fdc9ac46b2b9bd302606023ac81e78679364de6890e80"} Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.159717 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f9tj4" event={"ID":"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8","Type":"ContainerStarted","Data":"d8f4d951ff65d4bba3e13d8ce501f3f885c7b016da568d288783d358264f6688"} Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.159806 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f9tj4" event={"ID":"bff7cbe3-e11d-478e-8ce8-1a03becd8ac8","Type":"ContainerStarted","Data":"bda40b82aafd24d343905b66baf4f0df6615bf22eaf0b9cd022d06af6e89d190"} Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.165266 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-krnzn" podStartSLOduration=9.469723133 podStartE2EDuration="18.165240297s" podCreationTimestamp="2025-10-02 07:31:35 +0000 UTC" firstStartedPulling="2025-10-02 07:31:40.679440523 +0000 UTC m=+921.711386810" lastFinishedPulling="2025-10-02 07:31:49.374957687 +0000 UTC m=+930.406903974" observedRunningTime="2025-10-02 07:31:53.160754335 +0000 UTC m=+934.192700642" watchObservedRunningTime="2025-10-02 07:31:53.165240297 +0000 UTC m=+934.197186584" Oct 02 07:31:53 crc kubenswrapper[4960]: I1002 07:31:53.223946 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-f9tj4" podStartSLOduration=10.533636442 podStartE2EDuration="18.22391321s" podCreationTimestamp="2025-10-02 07:31:35 +0000 UTC" firstStartedPulling="2025-10-02 07:31:41.289393619 +0000 UTC m=+922.321339906" lastFinishedPulling="2025-10-02 07:31:48.979670387 +0000 UTC m=+930.011616674" observedRunningTime="2025-10-02 07:31:53.210668888 +0000 UTC m=+934.242615175" watchObservedRunningTime="2025-10-02 07:31:53.22391321 +0000 UTC m=+934.255859517" Oct 02 07:31:54 crc kubenswrapper[4960]: I1002 07:31:54.173610 4960 generic.go:334] "Generic (PLEG): container finished" podID="3be247d9-d9ed-49f1-9013-2ad50b6e4ed8" containerID="57ea5128660cbd00ec1430e2a3f0cd1755991e3ffa00e5722806d0ba687709c4" exitCode=0 Oct 02 07:31:54 crc kubenswrapper[4960]: I1002 07:31:54.173874 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8","Type":"ContainerDied","Data":"57ea5128660cbd00ec1430e2a3f0cd1755991e3ffa00e5722806d0ba687709c4"} Oct 02 07:31:54 crc kubenswrapper[4960]: I1002 07:31:54.174471 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:54 crc kubenswrapper[4960]: I1002 07:31:54.174516 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:31:54 crc kubenswrapper[4960]: I1002 07:31:54.342323 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" path="/var/lib/kubelet/pods/b847ef95-9ec1-44ae-b395-f1acd7bbea62/volumes" Oct 02 07:31:54 crc kubenswrapper[4960]: I1002 07:31:54.851779 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.187414 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"225801bc-ca07-4fa8-8d5c-1d56cfef957b","Type":"ContainerStarted","Data":"e60a2c776c8165acd26dc9dd8523f92e6ac3e5fbd818e47f32d0de2b1a5606f7"} Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.196960 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3be247d9-d9ed-49f1-9013-2ad50b6e4ed8","Type":"ContainerStarted","Data":"967e2037c2af999b36e93b2f431f03aa8fe40af74aae698da69092bee4ba95c9"} Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.206611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ea27c7c-a834-41de-9a07-3f0611c3bc23","Type":"ContainerStarted","Data":"bc5e41e4e179b2aa14798537349a1bce27be35e9a702816373e5d4601481781b"} Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.213635 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"67a98bd9-8fba-44c5-9229-de8a1053d9d6","Type":"ContainerStarted","Data":"86fa203f4f271205219d8355689c2abd96a3ae401299da0d4962a9426b2c2319"} Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.220458 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.573368928 podStartE2EDuration="21.22043647s" podCreationTimestamp="2025-10-02 07:31:34 +0000 UTC" firstStartedPulling="2025-10-02 07:31:40.754810193 +0000 UTC m=+921.786756480" lastFinishedPulling="2025-10-02 07:31:54.401877735 +0000 UTC m=+935.433824022" observedRunningTime="2025-10-02 07:31:55.217406198 +0000 UTC m=+936.249352485" watchObservedRunningTime="2025-10-02 07:31:55.22043647 +0000 UTC m=+936.252382757" Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.255729 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.721095145 podStartE2EDuration="18.255703624s" podCreationTimestamp="2025-10-02 07:31:37 +0000 UTC" firstStartedPulling="2025-10-02 07:31:40.855952147 +0000 UTC m=+921.887898444" lastFinishedPulling="2025-10-02 07:31:54.390560606 +0000 UTC m=+935.422506923" observedRunningTime="2025-10-02 07:31:55.24936349 +0000 UTC m=+936.281309797" watchObservedRunningTime="2025-10-02 07:31:55.255703624 +0000 UTC m=+936.287649911" Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.292251 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.627723383 podStartE2EDuration="28.292219991s" podCreationTimestamp="2025-10-02 07:31:27 +0000 UTC" firstStartedPulling="2025-10-02 07:31:38.579375985 +0000 UTC m=+919.611322312" lastFinishedPulling="2025-10-02 07:31:48.243872633 +0000 UTC m=+929.275818920" observedRunningTime="2025-10-02 07:31:55.281811087 +0000 UTC m=+936.313757384" watchObservedRunningTime="2025-10-02 07:31:55.292219991 +0000 UTC m=+936.324166298" Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.311538 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.231368819 podStartE2EDuration="28.311509309s" podCreationTimestamp="2025-10-02 07:31:27 +0000 UTC" firstStartedPulling="2025-10-02 07:31:40.215294812 +0000 UTC m=+921.247241099" lastFinishedPulling="2025-10-02 07:31:48.295435292 +0000 UTC m=+929.327381589" observedRunningTime="2025-10-02 07:31:55.307130849 +0000 UTC m=+936.339077146" watchObservedRunningTime="2025-10-02 07:31:55.311509309 +0000 UTC m=+936.343455616" Oct 02 07:31:55 crc kubenswrapper[4960]: I1002 07:31:55.600051 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:56 crc kubenswrapper[4960]: I1002 07:31:56.600038 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:56 crc kubenswrapper[4960]: I1002 07:31:56.642373 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.311159 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.329353 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.368260 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.607323 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f98bbbbf7-cg82t"] Oct 02 07:31:57 crc kubenswrapper[4960]: E1002 07:31:57.608055 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerName="init" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.608068 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerName="init" Oct 02 07:31:57 crc kubenswrapper[4960]: E1002 07:31:57.608076 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerName="dnsmasq-dns" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.608083 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerName="dnsmasq-dns" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.608263 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b847ef95-9ec1-44ae-b395-f1acd7bbea62" containerName="dnsmasq-dns" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.609146 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.612970 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.625589 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f98bbbbf7-cg82t"] Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.657904 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-hmxsf"] Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.659475 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.662545 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.685575 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hmxsf"] Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.685937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-ovsdbserver-nb\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.686034 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-dns-svc\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.686092 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-config\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.686122 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s94d6\" (UniqueName: \"kubernetes.io/projected/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-kube-api-access-s94d6\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.787865 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj485\" (UniqueName: \"kubernetes.io/projected/09ac27ea-acdf-444a-b8b8-0f61332416b1-kube-api-access-rj485\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788029 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-ovsdbserver-nb\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788079 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09ac27ea-acdf-444a-b8b8-0f61332416b1-ovs-rundir\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788103 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-dns-svc\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788123 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac27ea-acdf-444a-b8b8-0f61332416b1-config\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-config\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788162 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s94d6\" (UniqueName: \"kubernetes.io/projected/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-kube-api-access-s94d6\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788427 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ac27ea-acdf-444a-b8b8-0f61332416b1-combined-ca-bundle\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788525 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09ac27ea-acdf-444a-b8b8-0f61332416b1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.788652 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09ac27ea-acdf-444a-b8b8-0f61332416b1-ovn-rundir\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.789467 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-dns-svc\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.790137 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-ovsdbserver-nb\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.790216 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-config\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.813542 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s94d6\" (UniqueName: \"kubernetes.io/projected/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-kube-api-access-s94d6\") pod \"dnsmasq-dns-f98bbbbf7-cg82t\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.890331 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09ac27ea-acdf-444a-b8b8-0f61332416b1-ovn-rundir\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.890467 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj485\" (UniqueName: \"kubernetes.io/projected/09ac27ea-acdf-444a-b8b8-0f61332416b1-kube-api-access-rj485\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.890572 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09ac27ea-acdf-444a-b8b8-0f61332416b1-ovs-rundir\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.890614 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac27ea-acdf-444a-b8b8-0f61332416b1-config\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.890684 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ac27ea-acdf-444a-b8b8-0f61332416b1-combined-ca-bundle\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.890733 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09ac27ea-acdf-444a-b8b8-0f61332416b1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.892335 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09ac27ea-acdf-444a-b8b8-0f61332416b1-ovn-rundir\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.892791 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09ac27ea-acdf-444a-b8b8-0f61332416b1-ovs-rundir\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.893325 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09ac27ea-acdf-444a-b8b8-0f61332416b1-config\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.899709 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09ac27ea-acdf-444a-b8b8-0f61332416b1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.910217 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ac27ea-acdf-444a-b8b8-0f61332416b1-combined-ca-bundle\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.910577 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj485\" (UniqueName: \"kubernetes.io/projected/09ac27ea-acdf-444a-b8b8-0f61332416b1-kube-api-access-rj485\") pod \"ovn-controller-metrics-hmxsf\" (UID: \"09ac27ea-acdf-444a-b8b8-0f61332416b1\") " pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.933649 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.947734 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f98bbbbf7-cg82t"] Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.958856 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64dd86f89c-bppxm"] Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.963163 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.969252 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.987550 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dd86f89c-bppxm"] Oct 02 07:31:57 crc kubenswrapper[4960]: I1002 07:31:57.993192 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hmxsf" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.097237 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-sb\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.097293 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-config\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.097327 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5mzx\" (UniqueName: \"kubernetes.io/projected/a13a796e-8fc2-463c-9587-8fc19145de29-kube-api-access-s5mzx\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.097389 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-dns-svc\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.097438 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.198852 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-dns-svc\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.198935 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.199018 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-sb\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.199046 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-config\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.199072 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5mzx\" (UniqueName: \"kubernetes.io/projected/a13a796e-8fc2-463c-9587-8fc19145de29-kube-api-access-s5mzx\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.200064 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-dns-svc\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.200151 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-sb\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.200318 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.200476 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-config\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.224298 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5mzx\" (UniqueName: \"kubernetes.io/projected/a13a796e-8fc2-463c-9587-8fc19145de29-kube-api-access-s5mzx\") pod \"dnsmasq-dns-64dd86f89c-bppxm\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.244611 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.291778 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.355335 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.482481 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f98bbbbf7-cg82t"] Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.533221 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hmxsf"] Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.542572 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.544162 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.547542 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.547837 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.548247 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.549146 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7h5bm" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.558802 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.606057 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfrcx\" (UniqueName: \"kubernetes.io/projected/419dd9e4-dbbd-4a58-823b-b987afdf20cb-kube-api-access-dfrcx\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.606130 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/419dd9e4-dbbd-4a58-823b-b987afdf20cb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.606150 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.606170 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419dd9e4-dbbd-4a58-823b-b987afdf20cb-config\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.606194 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.606247 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.606272 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/419dd9e4-dbbd-4a58-823b-b987afdf20cb-scripts\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.713203 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfrcx\" (UniqueName: \"kubernetes.io/projected/419dd9e4-dbbd-4a58-823b-b987afdf20cb-kube-api-access-dfrcx\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.713588 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/419dd9e4-dbbd-4a58-823b-b987afdf20cb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.713612 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.713634 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419dd9e4-dbbd-4a58-823b-b987afdf20cb-config\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.713659 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.713721 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.713745 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/419dd9e4-dbbd-4a58-823b-b987afdf20cb-scripts\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.714058 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/419dd9e4-dbbd-4a58-823b-b987afdf20cb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.718344 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.718443 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.718759 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/419dd9e4-dbbd-4a58-823b-b987afdf20cb-scripts\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.718938 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419dd9e4-dbbd-4a58-823b-b987afdf20cb-config\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.726147 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419dd9e4-dbbd-4a58-823b-b987afdf20cb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.746538 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfrcx\" (UniqueName: \"kubernetes.io/projected/419dd9e4-dbbd-4a58-823b-b987afdf20cb-kube-api-access-dfrcx\") pod \"ovn-northd-0\" (UID: \"419dd9e4-dbbd-4a58-823b-b987afdf20cb\") " pod="openstack/ovn-northd-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.790196 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.791040 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 07:31:58 crc kubenswrapper[4960]: I1002 07:31:58.894167 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.077096 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dd86f89c-bppxm"] Oct 02 07:31:59 crc kubenswrapper[4960]: W1002 07:31:59.082224 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda13a796e_8fc2_463c_9587_8fc19145de29.slice/crio-e22d9fc4fde2defb3bc57de0f24c6f9cb47e705814fa88a65923bf210ebf8532 WatchSource:0}: Error finding container e22d9fc4fde2defb3bc57de0f24c6f9cb47e705814fa88a65923bf210ebf8532: Status 404 returned error can't find the container with id e22d9fc4fde2defb3bc57de0f24c6f9cb47e705814fa88a65923bf210ebf8532 Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.156053 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.156125 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.187712 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.188776 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.257133 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hmxsf" event={"ID":"09ac27ea-acdf-444a-b8b8-0f61332416b1","Type":"ContainerStarted","Data":"fafb67e260454f94687209fd7de76a1f4b43d02c4fc74cc608bf8633af241fbc"} Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.257194 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hmxsf" event={"ID":"09ac27ea-acdf-444a-b8b8-0f61332416b1","Type":"ContainerStarted","Data":"089d12cf64d8c92ff669789024b9af8206e389b0635b619509a27c1e77933e1a"} Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.259830 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" event={"ID":"a13a796e-8fc2-463c-9587-8fc19145de29","Type":"ContainerStarted","Data":"e22d9fc4fde2defb3bc57de0f24c6f9cb47e705814fa88a65923bf210ebf8532"} Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.262128 4960 generic.go:334] "Generic (PLEG): container finished" podID="4ddadf7d-e902-4070-8df7-7cd8d3e84aff" containerID="0aacbed6841cc4512da22bf78e6b79f69ac735dfc60ec3cdf5fbd8a3410a1fd2" exitCode=0 Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.263631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" event={"ID":"4ddadf7d-e902-4070-8df7-7cd8d3e84aff","Type":"ContainerDied","Data":"0aacbed6841cc4512da22bf78e6b79f69ac735dfc60ec3cdf5fbd8a3410a1fd2"} Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.263686 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" event={"ID":"4ddadf7d-e902-4070-8df7-7cd8d3e84aff","Type":"ContainerStarted","Data":"5b073e45097a7233566ee5c2c85d6ea9cf15a4af9e7e79af724b0d3de2043c16"} Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.289208 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-hmxsf" podStartSLOduration=2.289174269 podStartE2EDuration="2.289174269s" podCreationTimestamp="2025-10-02 07:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:31:59.279655568 +0000 UTC m=+940.311601885" watchObservedRunningTime="2025-10-02 07:31:59.289174269 +0000 UTC m=+940.321120586" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.359688 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.574688 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.735313 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-dns-svc\") pod \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.735378 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s94d6\" (UniqueName: \"kubernetes.io/projected/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-kube-api-access-s94d6\") pod \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.735446 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-ovsdbserver-nb\") pod \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.735677 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-config\") pod \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\" (UID: \"4ddadf7d-e902-4070-8df7-7cd8d3e84aff\") " Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.743018 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-kube-api-access-s94d6" (OuterVolumeSpecName: "kube-api-access-s94d6") pod "4ddadf7d-e902-4070-8df7-7cd8d3e84aff" (UID: "4ddadf7d-e902-4070-8df7-7cd8d3e84aff"). InnerVolumeSpecName "kube-api-access-s94d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.754895 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-config" (OuterVolumeSpecName: "config") pod "4ddadf7d-e902-4070-8df7-7cd8d3e84aff" (UID: "4ddadf7d-e902-4070-8df7-7cd8d3e84aff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.756835 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ddadf7d-e902-4070-8df7-7cd8d3e84aff" (UID: "4ddadf7d-e902-4070-8df7-7cd8d3e84aff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.781108 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ddadf7d-e902-4070-8df7-7cd8d3e84aff" (UID: "4ddadf7d-e902-4070-8df7-7cd8d3e84aff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.838648 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.838698 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.838725 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:59 crc kubenswrapper[4960]: I1002 07:31:59.838743 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s94d6\" (UniqueName: \"kubernetes.io/projected/4ddadf7d-e902-4070-8df7-7cd8d3e84aff-kube-api-access-s94d6\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.274650 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" event={"ID":"4ddadf7d-e902-4070-8df7-7cd8d3e84aff","Type":"ContainerDied","Data":"5b073e45097a7233566ee5c2c85d6ea9cf15a4af9e7e79af724b0d3de2043c16"} Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.274716 4960 scope.go:117] "RemoveContainer" containerID="0aacbed6841cc4512da22bf78e6b79f69ac735dfc60ec3cdf5fbd8a3410a1fd2" Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.274874 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f98bbbbf7-cg82t" Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.278800 4960 generic.go:334] "Generic (PLEG): container finished" podID="a13a796e-8fc2-463c-9587-8fc19145de29" containerID="c690443620a87e2afc239b0c63ed3e580d9860c8474aa2074a74e3008efb379a" exitCode=0 Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.278901 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" event={"ID":"a13a796e-8fc2-463c-9587-8fc19145de29","Type":"ContainerDied","Data":"c690443620a87e2afc239b0c63ed3e580d9860c8474aa2074a74e3008efb379a"} Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.284958 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"419dd9e4-dbbd-4a58-823b-b987afdf20cb","Type":"ContainerStarted","Data":"fe44b2b0eb7592d0e0d719d0f3d4d4495a2102a9b49428ffdf667aed95a35f98"} Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.420509 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f98bbbbf7-cg82t"] Oct 02 07:32:00 crc kubenswrapper[4960]: I1002 07:32:00.421946 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f98bbbbf7-cg82t"] Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.294078 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" event={"ID":"a13a796e-8fc2-463c-9587-8fc19145de29","Type":"ContainerStarted","Data":"89b90dec3c5e437f50f3c08f10cc2f3aef20852c15382cb33db6ad69795db39a"} Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.294505 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.297187 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"419dd9e4-dbbd-4a58-823b-b987afdf20cb","Type":"ContainerStarted","Data":"8d8945ccca0ced1f755474f2dcd29b6008c9af47b2aa84657f23dd6eeabad27e"} Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.297295 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.297313 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"419dd9e4-dbbd-4a58-823b-b987afdf20cb","Type":"ContainerStarted","Data":"e1b8390dcfbc9d94405ced41682e5d8938edbe5340805d0ca55edb634e3c1f2f"} Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.317064 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" podStartSLOduration=4.317029424 podStartE2EDuration="4.317029424s" podCreationTimestamp="2025-10-02 07:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:32:01.313388304 +0000 UTC m=+942.345334601" watchObservedRunningTime="2025-10-02 07:32:01.317029424 +0000 UTC m=+942.348975721" Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.345509 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.051187849 podStartE2EDuration="3.345492322s" podCreationTimestamp="2025-10-02 07:31:58 +0000 UTC" firstStartedPulling="2025-10-02 07:31:59.445053297 +0000 UTC m=+940.476999594" lastFinishedPulling="2025-10-02 07:32:00.73935777 +0000 UTC m=+941.771304067" observedRunningTime="2025-10-02 07:32:01.341654588 +0000 UTC m=+942.373600885" watchObservedRunningTime="2025-10-02 07:32:01.345492322 +0000 UTC m=+942.377438599" Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.372505 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.434601 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 07:32:01 crc kubenswrapper[4960]: I1002 07:32:01.593369 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 07:32:02 crc kubenswrapper[4960]: I1002 07:32:02.346680 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ddadf7d-e902-4070-8df7-7cd8d3e84aff" path="/var/lib/kubelet/pods/4ddadf7d-e902-4070-8df7-7cd8d3e84aff/volumes" Oct 02 07:32:03 crc kubenswrapper[4960]: I1002 07:32:03.259905 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:03 crc kubenswrapper[4960]: I1002 07:32:03.322836 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.179566 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-sck8h"] Oct 02 07:32:05 crc kubenswrapper[4960]: E1002 07:32:05.181028 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddadf7d-e902-4070-8df7-7cd8d3e84aff" containerName="init" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.181058 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddadf7d-e902-4070-8df7-7cd8d3e84aff" containerName="init" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.181383 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddadf7d-e902-4070-8df7-7cd8d3e84aff" containerName="init" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.182377 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sck8h" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.205888 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sck8h"] Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.264916 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qgbt\" (UniqueName: \"kubernetes.io/projected/2a808721-6be5-4339-b4cc-770c6f288564-kube-api-access-4qgbt\") pod \"glance-db-create-sck8h\" (UID: \"2a808721-6be5-4339-b4cc-770c6f288564\") " pod="openstack/glance-db-create-sck8h" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.367333 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qgbt\" (UniqueName: \"kubernetes.io/projected/2a808721-6be5-4339-b4cc-770c6f288564-kube-api-access-4qgbt\") pod \"glance-db-create-sck8h\" (UID: \"2a808721-6be5-4339-b4cc-770c6f288564\") " pod="openstack/glance-db-create-sck8h" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.405277 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qgbt\" (UniqueName: \"kubernetes.io/projected/2a808721-6be5-4339-b4cc-770c6f288564-kube-api-access-4qgbt\") pod \"glance-db-create-sck8h\" (UID: \"2a808721-6be5-4339-b4cc-770c6f288564\") " pod="openstack/glance-db-create-sck8h" Oct 02 07:32:05 crc kubenswrapper[4960]: I1002 07:32:05.508420 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sck8h" Oct 02 07:32:06 crc kubenswrapper[4960]: I1002 07:32:06.038782 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sck8h"] Oct 02 07:32:06 crc kubenswrapper[4960]: I1002 07:32:06.354764 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sck8h" event={"ID":"2a808721-6be5-4339-b4cc-770c6f288564","Type":"ContainerStarted","Data":"c55ad313c401ecf9edf935b96cfc608b1cd48b63bc87ea5427af8d43c771cfe9"} Oct 02 07:32:06 crc kubenswrapper[4960]: I1002 07:32:06.354815 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sck8h" event={"ID":"2a808721-6be5-4339-b4cc-770c6f288564","Type":"ContainerStarted","Data":"804ef5930d6496e353e32c86a8579ed73e1828060c912d138ec86ae32975fea4"} Oct 02 07:32:06 crc kubenswrapper[4960]: I1002 07:32:06.384011 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-sck8h" podStartSLOduration=1.383991227 podStartE2EDuration="1.383991227s" podCreationTimestamp="2025-10-02 07:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:32:06.379546115 +0000 UTC m=+947.411492422" watchObservedRunningTime="2025-10-02 07:32:06.383991227 +0000 UTC m=+947.415937524" Oct 02 07:32:07 crc kubenswrapper[4960]: I1002 07:32:07.367734 4960 generic.go:334] "Generic (PLEG): container finished" podID="2a808721-6be5-4339-b4cc-770c6f288564" containerID="c55ad313c401ecf9edf935b96cfc608b1cd48b63bc87ea5427af8d43c771cfe9" exitCode=0 Oct 02 07:32:07 crc kubenswrapper[4960]: I1002 07:32:07.367848 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sck8h" event={"ID":"2a808721-6be5-4339-b4cc-770c6f288564","Type":"ContainerDied","Data":"c55ad313c401ecf9edf935b96cfc608b1cd48b63bc87ea5427af8d43c771cfe9"} Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.358283 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.435349 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-qnq6h"] Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.435693 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" podUID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerName="dnsmasq-dns" containerID="cri-o://67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a" gracePeriod=10 Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.824673 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sck8h" Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.916248 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.950445 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-config\") pod \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.950500 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64tfm\" (UniqueName: \"kubernetes.io/projected/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-kube-api-access-64tfm\") pod \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.950563 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qgbt\" (UniqueName: \"kubernetes.io/projected/2a808721-6be5-4339-b4cc-770c6f288564-kube-api-access-4qgbt\") pod \"2a808721-6be5-4339-b4cc-770c6f288564\" (UID: \"2a808721-6be5-4339-b4cc-770c6f288564\") " Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.950591 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-dns-svc\") pod \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\" (UID: \"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e\") " Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.957560 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a808721-6be5-4339-b4cc-770c6f288564-kube-api-access-4qgbt" (OuterVolumeSpecName: "kube-api-access-4qgbt") pod "2a808721-6be5-4339-b4cc-770c6f288564" (UID: "2a808721-6be5-4339-b4cc-770c6f288564"). InnerVolumeSpecName "kube-api-access-4qgbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.958343 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-kube-api-access-64tfm" (OuterVolumeSpecName: "kube-api-access-64tfm") pod "845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" (UID: "845da1dc-0705-4bfa-b0f4-dbce9e6cda6e"). InnerVolumeSpecName "kube-api-access-64tfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:08 crc kubenswrapper[4960]: I1002 07:32:08.991835 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-config" (OuterVolumeSpecName: "config") pod "845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" (UID: "845da1dc-0705-4bfa-b0f4-dbce9e6cda6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.006167 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" (UID: "845da1dc-0705-4bfa-b0f4-dbce9e6cda6e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.053563 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.053602 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64tfm\" (UniqueName: \"kubernetes.io/projected/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-kube-api-access-64tfm\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.053620 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qgbt\" (UniqueName: \"kubernetes.io/projected/2a808721-6be5-4339-b4cc-770c6f288564-kube-api-access-4qgbt\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.053633 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.386823 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sck8h" event={"ID":"2a808721-6be5-4339-b4cc-770c6f288564","Type":"ContainerDied","Data":"804ef5930d6496e353e32c86a8579ed73e1828060c912d138ec86ae32975fea4"} Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.386865 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="804ef5930d6496e353e32c86a8579ed73e1828060c912d138ec86ae32975fea4" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.387328 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sck8h" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.388912 4960 generic.go:334] "Generic (PLEG): container finished" podID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerID="67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a" exitCode=0 Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.388952 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" event={"ID":"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e","Type":"ContainerDied","Data":"67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a"} Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.389001 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" event={"ID":"845da1dc-0705-4bfa-b0f4-dbce9e6cda6e","Type":"ContainerDied","Data":"7fad62d309c49af1558e4a198083b680a766d4650571dcb0ca5c94401fe61ae3"} Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.389013 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-qnq6h" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.389020 4960 scope.go:117] "RemoveContainer" containerID="67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.435588 4960 scope.go:117] "RemoveContainer" containerID="4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.440869 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-qnq6h"] Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.451610 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-qnq6h"] Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.468228 4960 scope.go:117] "RemoveContainer" containerID="67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a" Oct 02 07:32:09 crc kubenswrapper[4960]: E1002 07:32:09.474575 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a\": container with ID starting with 67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a not found: ID does not exist" containerID="67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.474706 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a"} err="failed to get container status \"67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a\": rpc error: code = NotFound desc = could not find container \"67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a\": container with ID starting with 67a9b53556803ad8f3e26d0f464e5c97c7706dd92f61c5bce61f401739b4107a not found: ID does not exist" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.474821 4960 scope.go:117] "RemoveContainer" containerID="4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6" Oct 02 07:32:09 crc kubenswrapper[4960]: E1002 07:32:09.475698 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6\": container with ID starting with 4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6 not found: ID does not exist" containerID="4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.475904 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6"} err="failed to get container status \"4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6\": rpc error: code = NotFound desc = could not find container \"4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6\": container with ID starting with 4aad9a97ccbd68d8c18bd3a46c0bd3f914e383fef1c80c8a6d172d41a9f74aa6 not found: ID does not exist" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.492895 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-xr5gr"] Oct 02 07:32:09 crc kubenswrapper[4960]: E1002 07:32:09.493694 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a808721-6be5-4339-b4cc-770c6f288564" containerName="mariadb-database-create" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.493737 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a808721-6be5-4339-b4cc-770c6f288564" containerName="mariadb-database-create" Oct 02 07:32:09 crc kubenswrapper[4960]: E1002 07:32:09.493762 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerName="init" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.493771 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerName="init" Oct 02 07:32:09 crc kubenswrapper[4960]: E1002 07:32:09.493799 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerName="dnsmasq-dns" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.493806 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerName="dnsmasq-dns" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.496329 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a808721-6be5-4339-b4cc-770c6f288564" containerName="mariadb-database-create" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.496351 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" containerName="dnsmasq-dns" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.497266 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xr5gr"] Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.497369 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xr5gr" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.576173 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75jff\" (UniqueName: \"kubernetes.io/projected/b98a8729-a22e-49e4-ae15-6ff17fb6c4b7-kube-api-access-75jff\") pod \"keystone-db-create-xr5gr\" (UID: \"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7\") " pod="openstack/keystone-db-create-xr5gr" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.679149 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75jff\" (UniqueName: \"kubernetes.io/projected/b98a8729-a22e-49e4-ae15-6ff17fb6c4b7-kube-api-access-75jff\") pod \"keystone-db-create-xr5gr\" (UID: \"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7\") " pod="openstack/keystone-db-create-xr5gr" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.703522 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75jff\" (UniqueName: \"kubernetes.io/projected/b98a8729-a22e-49e4-ae15-6ff17fb6c4b7-kube-api-access-75jff\") pod \"keystone-db-create-xr5gr\" (UID: \"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7\") " pod="openstack/keystone-db-create-xr5gr" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.795438 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-p2r6q"] Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.797263 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p2r6q" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.811452 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-p2r6q"] Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.853477 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xr5gr" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.882734 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9czf2\" (UniqueName: \"kubernetes.io/projected/109e5c0f-e848-4762-bdc8-1ae57e4f7fd7-kube-api-access-9czf2\") pod \"placement-db-create-p2r6q\" (UID: \"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7\") " pod="openstack/placement-db-create-p2r6q" Oct 02 07:32:09 crc kubenswrapper[4960]: I1002 07:32:09.985534 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9czf2\" (UniqueName: \"kubernetes.io/projected/109e5c0f-e848-4762-bdc8-1ae57e4f7fd7-kube-api-access-9czf2\") pod \"placement-db-create-p2r6q\" (UID: \"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7\") " pod="openstack/placement-db-create-p2r6q" Oct 02 07:32:10 crc kubenswrapper[4960]: I1002 07:32:10.017218 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9czf2\" (UniqueName: \"kubernetes.io/projected/109e5c0f-e848-4762-bdc8-1ae57e4f7fd7-kube-api-access-9czf2\") pod \"placement-db-create-p2r6q\" (UID: \"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7\") " pod="openstack/placement-db-create-p2r6q" Oct 02 07:32:10 crc kubenswrapper[4960]: I1002 07:32:10.113226 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xr5gr"] Oct 02 07:32:10 crc kubenswrapper[4960]: I1002 07:32:10.117690 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p2r6q" Oct 02 07:32:10 crc kubenswrapper[4960]: I1002 07:32:10.347054 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="845da1dc-0705-4bfa-b0f4-dbce9e6cda6e" path="/var/lib/kubelet/pods/845da1dc-0705-4bfa-b0f4-dbce9e6cda6e/volumes" Oct 02 07:32:10 crc kubenswrapper[4960]: I1002 07:32:10.411766 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xr5gr" event={"ID":"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7","Type":"ContainerStarted","Data":"86a0aeb86d65165f2bdaa438b63c62cc70b9f17c07ae8704e79d8c16d1ec0560"} Oct 02 07:32:10 crc kubenswrapper[4960]: I1002 07:32:10.411823 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xr5gr" event={"ID":"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7","Type":"ContainerStarted","Data":"0c4d34ea35efaa36fa3301364e74216822c48746c27b7acb48f3511890cb5c15"} Oct 02 07:32:10 crc kubenswrapper[4960]: I1002 07:32:10.413517 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-p2r6q"] Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.422475 4960 generic.go:334] "Generic (PLEG): container finished" podID="b98a8729-a22e-49e4-ae15-6ff17fb6c4b7" containerID="86a0aeb86d65165f2bdaa438b63c62cc70b9f17c07ae8704e79d8c16d1ec0560" exitCode=0 Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.422577 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xr5gr" event={"ID":"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7","Type":"ContainerDied","Data":"86a0aeb86d65165f2bdaa438b63c62cc70b9f17c07ae8704e79d8c16d1ec0560"} Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.424696 4960 generic.go:334] "Generic (PLEG): container finished" podID="109e5c0f-e848-4762-bdc8-1ae57e4f7fd7" containerID="7cb50b5d882242ae41e7d94e452a1d1fad0925228252dcea8f75ed4a513110c9" exitCode=0 Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.424743 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p2r6q" event={"ID":"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7","Type":"ContainerDied","Data":"7cb50b5d882242ae41e7d94e452a1d1fad0925228252dcea8f75ed4a513110c9"} Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.424762 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p2r6q" event={"ID":"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7","Type":"ContainerStarted","Data":"f31ac5394d877abd358ad8ee10130f6f068fef6ba86ac67576cc07e34a21fd02"} Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.811841 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xr5gr" Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.929175 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75jff\" (UniqueName: \"kubernetes.io/projected/b98a8729-a22e-49e4-ae15-6ff17fb6c4b7-kube-api-access-75jff\") pod \"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7\" (UID: \"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7\") " Oct 02 07:32:11 crc kubenswrapper[4960]: I1002 07:32:11.938667 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b98a8729-a22e-49e4-ae15-6ff17fb6c4b7-kube-api-access-75jff" (OuterVolumeSpecName: "kube-api-access-75jff") pod "b98a8729-a22e-49e4-ae15-6ff17fb6c4b7" (UID: "b98a8729-a22e-49e4-ae15-6ff17fb6c4b7"). InnerVolumeSpecName "kube-api-access-75jff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.031099 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75jff\" (UniqueName: \"kubernetes.io/projected/b98a8729-a22e-49e4-ae15-6ff17fb6c4b7-kube-api-access-75jff\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.441451 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xr5gr" Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.442917 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xr5gr" event={"ID":"b98a8729-a22e-49e4-ae15-6ff17fb6c4b7","Type":"ContainerDied","Data":"0c4d34ea35efaa36fa3301364e74216822c48746c27b7acb48f3511890cb5c15"} Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.442953 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c4d34ea35efaa36fa3301364e74216822c48746c27b7acb48f3511890cb5c15" Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.765209 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p2r6q" Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.846680 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9czf2\" (UniqueName: \"kubernetes.io/projected/109e5c0f-e848-4762-bdc8-1ae57e4f7fd7-kube-api-access-9czf2\") pod \"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7\" (UID: \"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7\") " Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.858593 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/109e5c0f-e848-4762-bdc8-1ae57e4f7fd7-kube-api-access-9czf2" (OuterVolumeSpecName: "kube-api-access-9czf2") pod "109e5c0f-e848-4762-bdc8-1ae57e4f7fd7" (UID: "109e5c0f-e848-4762-bdc8-1ae57e4f7fd7"). InnerVolumeSpecName "kube-api-access-9czf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:12 crc kubenswrapper[4960]: I1002 07:32:12.950807 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9czf2\" (UniqueName: \"kubernetes.io/projected/109e5c0f-e848-4762-bdc8-1ae57e4f7fd7-kube-api-access-9czf2\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:13 crc kubenswrapper[4960]: I1002 07:32:13.452696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p2r6q" event={"ID":"109e5c0f-e848-4762-bdc8-1ae57e4f7fd7","Type":"ContainerDied","Data":"f31ac5394d877abd358ad8ee10130f6f068fef6ba86ac67576cc07e34a21fd02"} Oct 02 07:32:13 crc kubenswrapper[4960]: I1002 07:32:13.453319 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f31ac5394d877abd358ad8ee10130f6f068fef6ba86ac67576cc07e34a21fd02" Oct 02 07:32:13 crc kubenswrapper[4960]: I1002 07:32:13.452786 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p2r6q" Oct 02 07:32:13 crc kubenswrapper[4960]: I1002 07:32:13.996855 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.238711 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ef5b-account-create-sdwnj"] Oct 02 07:32:15 crc kubenswrapper[4960]: E1002 07:32:15.240087 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98a8729-a22e-49e4-ae15-6ff17fb6c4b7" containerName="mariadb-database-create" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.240116 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98a8729-a22e-49e4-ae15-6ff17fb6c4b7" containerName="mariadb-database-create" Oct 02 07:32:15 crc kubenswrapper[4960]: E1002 07:32:15.240171 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109e5c0f-e848-4762-bdc8-1ae57e4f7fd7" containerName="mariadb-database-create" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.240184 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="109e5c0f-e848-4762-bdc8-1ae57e4f7fd7" containerName="mariadb-database-create" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.240488 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98a8729-a22e-49e4-ae15-6ff17fb6c4b7" containerName="mariadb-database-create" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.240518 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="109e5c0f-e848-4762-bdc8-1ae57e4f7fd7" containerName="mariadb-database-create" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.241529 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ef5b-account-create-sdwnj" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.244549 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.250642 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ef5b-account-create-sdwnj"] Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.300184 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mt2g\" (UniqueName: \"kubernetes.io/projected/44887da1-2bcd-482d-a53e-75e1ed397f84-kube-api-access-9mt2g\") pod \"glance-ef5b-account-create-sdwnj\" (UID: \"44887da1-2bcd-482d-a53e-75e1ed397f84\") " pod="openstack/glance-ef5b-account-create-sdwnj" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.402510 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mt2g\" (UniqueName: \"kubernetes.io/projected/44887da1-2bcd-482d-a53e-75e1ed397f84-kube-api-access-9mt2g\") pod \"glance-ef5b-account-create-sdwnj\" (UID: \"44887da1-2bcd-482d-a53e-75e1ed397f84\") " pod="openstack/glance-ef5b-account-create-sdwnj" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.430908 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mt2g\" (UniqueName: \"kubernetes.io/projected/44887da1-2bcd-482d-a53e-75e1ed397f84-kube-api-access-9mt2g\") pod \"glance-ef5b-account-create-sdwnj\" (UID: \"44887da1-2bcd-482d-a53e-75e1ed397f84\") " pod="openstack/glance-ef5b-account-create-sdwnj" Oct 02 07:32:15 crc kubenswrapper[4960]: I1002 07:32:15.562418 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ef5b-account-create-sdwnj" Oct 02 07:32:16 crc kubenswrapper[4960]: I1002 07:32:16.175031 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ef5b-account-create-sdwnj"] Oct 02 07:32:16 crc kubenswrapper[4960]: I1002 07:32:16.487195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ef5b-account-create-sdwnj" event={"ID":"44887da1-2bcd-482d-a53e-75e1ed397f84","Type":"ContainerStarted","Data":"aacf31b34dfb947084057bc6b1d6e3a965ea5171f0bc592fb2a4fc44ff5dc802"} Oct 02 07:32:16 crc kubenswrapper[4960]: I1002 07:32:16.487270 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ef5b-account-create-sdwnj" event={"ID":"44887da1-2bcd-482d-a53e-75e1ed397f84","Type":"ContainerStarted","Data":"d0fbc9a7960520612d48c0a65a53f9ca72f7dc496496f905895a376b198f737b"} Oct 02 07:32:16 crc kubenswrapper[4960]: I1002 07:32:16.523944 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ef5b-account-create-sdwnj" podStartSLOduration=1.523907916 podStartE2EDuration="1.523907916s" podCreationTimestamp="2025-10-02 07:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:32:16.511463186 +0000 UTC m=+957.543409553" watchObservedRunningTime="2025-10-02 07:32:16.523907916 +0000 UTC m=+957.555854233" Oct 02 07:32:17 crc kubenswrapper[4960]: I1002 07:32:17.504803 4960 generic.go:334] "Generic (PLEG): container finished" podID="44887da1-2bcd-482d-a53e-75e1ed397f84" containerID="aacf31b34dfb947084057bc6b1d6e3a965ea5171f0bc592fb2a4fc44ff5dc802" exitCode=0 Oct 02 07:32:17 crc kubenswrapper[4960]: I1002 07:32:17.504933 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ef5b-account-create-sdwnj" event={"ID":"44887da1-2bcd-482d-a53e-75e1ed397f84","Type":"ContainerDied","Data":"aacf31b34dfb947084057bc6b1d6e3a965ea5171f0bc592fb2a4fc44ff5dc802"} Oct 02 07:32:18 crc kubenswrapper[4960]: I1002 07:32:18.883103 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ef5b-account-create-sdwnj" Oct 02 07:32:18 crc kubenswrapper[4960]: I1002 07:32:18.989235 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mt2g\" (UniqueName: \"kubernetes.io/projected/44887da1-2bcd-482d-a53e-75e1ed397f84-kube-api-access-9mt2g\") pod \"44887da1-2bcd-482d-a53e-75e1ed397f84\" (UID: \"44887da1-2bcd-482d-a53e-75e1ed397f84\") " Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.005631 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44887da1-2bcd-482d-a53e-75e1ed397f84-kube-api-access-9mt2g" (OuterVolumeSpecName: "kube-api-access-9mt2g") pod "44887da1-2bcd-482d-a53e-75e1ed397f84" (UID: "44887da1-2bcd-482d-a53e-75e1ed397f84"). InnerVolumeSpecName "kube-api-access-9mt2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.090899 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mt2g\" (UniqueName: \"kubernetes.io/projected/44887da1-2bcd-482d-a53e-75e1ed397f84-kube-api-access-9mt2g\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.531560 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ef5b-account-create-sdwnj" event={"ID":"44887da1-2bcd-482d-a53e-75e1ed397f84","Type":"ContainerDied","Data":"d0fbc9a7960520612d48c0a65a53f9ca72f7dc496496f905895a376b198f737b"} Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.531635 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0fbc9a7960520612d48c0a65a53f9ca72f7dc496496f905895a376b198f737b" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.531743 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ef5b-account-create-sdwnj" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.711644 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8399-account-create-r2ck2"] Oct 02 07:32:19 crc kubenswrapper[4960]: E1002 07:32:19.712782 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44887da1-2bcd-482d-a53e-75e1ed397f84" containerName="mariadb-account-create" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.712810 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="44887da1-2bcd-482d-a53e-75e1ed397f84" containerName="mariadb-account-create" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.713268 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="44887da1-2bcd-482d-a53e-75e1ed397f84" containerName="mariadb-account-create" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.715542 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8399-account-create-r2ck2" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.719424 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.722436 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8399-account-create-r2ck2"] Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.818765 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x654v\" (UniqueName: \"kubernetes.io/projected/3fb0af53-ac0a-43d0-a651-49f1970f13f7-kube-api-access-x654v\") pod \"keystone-8399-account-create-r2ck2\" (UID: \"3fb0af53-ac0a-43d0-a651-49f1970f13f7\") " pod="openstack/keystone-8399-account-create-r2ck2" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.921149 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x654v\" (UniqueName: \"kubernetes.io/projected/3fb0af53-ac0a-43d0-a651-49f1970f13f7-kube-api-access-x654v\") pod \"keystone-8399-account-create-r2ck2\" (UID: \"3fb0af53-ac0a-43d0-a651-49f1970f13f7\") " pod="openstack/keystone-8399-account-create-r2ck2" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.945954 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x654v\" (UniqueName: \"kubernetes.io/projected/3fb0af53-ac0a-43d0-a651-49f1970f13f7-kube-api-access-x654v\") pod \"keystone-8399-account-create-r2ck2\" (UID: \"3fb0af53-ac0a-43d0-a651-49f1970f13f7\") " pod="openstack/keystone-8399-account-create-r2ck2" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.986814 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8808-account-create-lk8m6"] Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.988336 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8808-account-create-lk8m6" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.991181 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 07:32:19 crc kubenswrapper[4960]: I1002 07:32:19.998833 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8808-account-create-lk8m6"] Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.039777 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8399-account-create-r2ck2" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.126017 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fzbh\" (UniqueName: \"kubernetes.io/projected/d33dff64-7c3f-404c-9eb4-5793c6bec19b-kube-api-access-5fzbh\") pod \"placement-8808-account-create-lk8m6\" (UID: \"d33dff64-7c3f-404c-9eb4-5793c6bec19b\") " pod="openstack/placement-8808-account-create-lk8m6" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.228422 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fzbh\" (UniqueName: \"kubernetes.io/projected/d33dff64-7c3f-404c-9eb4-5793c6bec19b-kube-api-access-5fzbh\") pod \"placement-8808-account-create-lk8m6\" (UID: \"d33dff64-7c3f-404c-9eb4-5793c6bec19b\") " pod="openstack/placement-8808-account-create-lk8m6" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.259073 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fzbh\" (UniqueName: \"kubernetes.io/projected/d33dff64-7c3f-404c-9eb4-5793c6bec19b-kube-api-access-5fzbh\") pod \"placement-8808-account-create-lk8m6\" (UID: \"d33dff64-7c3f-404c-9eb4-5793c6bec19b\") " pod="openstack/placement-8808-account-create-lk8m6" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.313161 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8808-account-create-lk8m6" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.406379 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-w875j"] Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.415346 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.419209 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4gqq6" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.419383 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.421216 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-w875j"] Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.544494 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8399-account-create-r2ck2"] Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.547805 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-787s4\" (UniqueName: \"kubernetes.io/projected/85344e88-9937-4cb2-903a-d62a423ce703-kube-api-access-787s4\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.547866 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-combined-ca-bundle\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.547901 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-config-data\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.548215 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-db-sync-config-data\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: W1002 07:32:20.553474 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fb0af53_ac0a_43d0_a651_49f1970f13f7.slice/crio-549d743905f8e48af0d2af4715cfe84178c753502af900f6620130a15483bbcc WatchSource:0}: Error finding container 549d743905f8e48af0d2af4715cfe84178c753502af900f6620130a15483bbcc: Status 404 returned error can't find the container with id 549d743905f8e48af0d2af4715cfe84178c753502af900f6620130a15483bbcc Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.559541 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.650297 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-787s4\" (UniqueName: \"kubernetes.io/projected/85344e88-9937-4cb2-903a-d62a423ce703-kube-api-access-787s4\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.650378 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-combined-ca-bundle\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.650419 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-config-data\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.650575 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-db-sync-config-data\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.658107 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-config-data\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.659165 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-db-sync-config-data\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.663136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-combined-ca-bundle\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.672135 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-787s4\" (UniqueName: \"kubernetes.io/projected/85344e88-9937-4cb2-903a-d62a423ce703-kube-api-access-787s4\") pod \"glance-db-sync-w875j\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.749698 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-w875j" Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.815854 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8808-account-create-lk8m6"] Oct 02 07:32:20 crc kubenswrapper[4960]: I1002 07:32:20.867131 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.101928 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-w875j"] Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.553863 4960 generic.go:334] "Generic (PLEG): container finished" podID="3fb0af53-ac0a-43d0-a651-49f1970f13f7" containerID="3226f9377160d51ee6cfdab42ebdfceeee9c1d29f98923450fcf0505b25cac57" exitCode=0 Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.553934 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8399-account-create-r2ck2" event={"ID":"3fb0af53-ac0a-43d0-a651-49f1970f13f7","Type":"ContainerDied","Data":"3226f9377160d51ee6cfdab42ebdfceeee9c1d29f98923450fcf0505b25cac57"} Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.554046 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8399-account-create-r2ck2" event={"ID":"3fb0af53-ac0a-43d0-a651-49f1970f13f7","Type":"ContainerStarted","Data":"549d743905f8e48af0d2af4715cfe84178c753502af900f6620130a15483bbcc"} Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.556716 4960 generic.go:334] "Generic (PLEG): container finished" podID="d33dff64-7c3f-404c-9eb4-5793c6bec19b" containerID="7416f347d2e27dbb509a6ad2ba9d46c21e6f28f8898f31098f0f2d07ee5ce2c2" exitCode=0 Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.556803 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8808-account-create-lk8m6" event={"ID":"d33dff64-7c3f-404c-9eb4-5793c6bec19b","Type":"ContainerDied","Data":"7416f347d2e27dbb509a6ad2ba9d46c21e6f28f8898f31098f0f2d07ee5ce2c2"} Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.556836 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8808-account-create-lk8m6" event={"ID":"d33dff64-7c3f-404c-9eb4-5793c6bec19b","Type":"ContainerStarted","Data":"50d340e31faa9cde7aa58efb57dfdfb080c3b8071906b7dd8b73cc7baf64b4a6"} Oct 02 07:32:21 crc kubenswrapper[4960]: I1002 07:32:21.559027 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-w875j" event={"ID":"85344e88-9937-4cb2-903a-d62a423ce703","Type":"ContainerStarted","Data":"28fb8f27a341fae9a04752e8fe961caef579eb6f43f0d9255e68a888e062e7fa"} Oct 02 07:32:22 crc kubenswrapper[4960]: I1002 07:32:22.988279 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8399-account-create-r2ck2" Oct 02 07:32:22 crc kubenswrapper[4960]: I1002 07:32:22.999091 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8808-account-create-lk8m6" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.098553 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x654v\" (UniqueName: \"kubernetes.io/projected/3fb0af53-ac0a-43d0-a651-49f1970f13f7-kube-api-access-x654v\") pod \"3fb0af53-ac0a-43d0-a651-49f1970f13f7\" (UID: \"3fb0af53-ac0a-43d0-a651-49f1970f13f7\") " Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.098654 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fzbh\" (UniqueName: \"kubernetes.io/projected/d33dff64-7c3f-404c-9eb4-5793c6bec19b-kube-api-access-5fzbh\") pod \"d33dff64-7c3f-404c-9eb4-5793c6bec19b\" (UID: \"d33dff64-7c3f-404c-9eb4-5793c6bec19b\") " Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.106771 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb0af53-ac0a-43d0-a651-49f1970f13f7-kube-api-access-x654v" (OuterVolumeSpecName: "kube-api-access-x654v") pod "3fb0af53-ac0a-43d0-a651-49f1970f13f7" (UID: "3fb0af53-ac0a-43d0-a651-49f1970f13f7"). InnerVolumeSpecName "kube-api-access-x654v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.107674 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d33dff64-7c3f-404c-9eb4-5793c6bec19b-kube-api-access-5fzbh" (OuterVolumeSpecName: "kube-api-access-5fzbh") pod "d33dff64-7c3f-404c-9eb4-5793c6bec19b" (UID: "d33dff64-7c3f-404c-9eb4-5793c6bec19b"). InnerVolumeSpecName "kube-api-access-5fzbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.200452 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x654v\" (UniqueName: \"kubernetes.io/projected/3fb0af53-ac0a-43d0-a651-49f1970f13f7-kube-api-access-x654v\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.200490 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fzbh\" (UniqueName: \"kubernetes.io/projected/d33dff64-7c3f-404c-9eb4-5793c6bec19b-kube-api-access-5fzbh\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.582842 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8399-account-create-r2ck2" event={"ID":"3fb0af53-ac0a-43d0-a651-49f1970f13f7","Type":"ContainerDied","Data":"549d743905f8e48af0d2af4715cfe84178c753502af900f6620130a15483bbcc"} Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.583309 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="549d743905f8e48af0d2af4715cfe84178c753502af900f6620130a15483bbcc" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.582867 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8399-account-create-r2ck2" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.585487 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8808-account-create-lk8m6" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.585518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8808-account-create-lk8m6" event={"ID":"d33dff64-7c3f-404c-9eb4-5793c6bec19b","Type":"ContainerDied","Data":"50d340e31faa9cde7aa58efb57dfdfb080c3b8071906b7dd8b73cc7baf64b4a6"} Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.585591 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50d340e31faa9cde7aa58efb57dfdfb080c3b8071906b7dd8b73cc7baf64b4a6" Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.588165 4960 generic.go:334] "Generic (PLEG): container finished" podID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerID="08856e4cd8439c757573e71be33fb3eca2f53d6f1f0a9f9e755f234506844f0f" exitCode=0 Oct 02 07:32:23 crc kubenswrapper[4960]: I1002 07:32:23.588225 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8c7a2f9-57bf-4520-b078-ce75a2e051ad","Type":"ContainerDied","Data":"08856e4cd8439c757573e71be33fb3eca2f53d6f1f0a9f9e755f234506844f0f"} Oct 02 07:32:24 crc kubenswrapper[4960]: I1002 07:32:24.599371 4960 generic.go:334] "Generic (PLEG): container finished" podID="36c68533-435a-4bb7-b338-4cd484b155f6" containerID="9bd820698bae00820e1521389d7e85c9f91579903f903da17b0bbb829e24c442" exitCode=0 Oct 02 07:32:24 crc kubenswrapper[4960]: I1002 07:32:24.599964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"36c68533-435a-4bb7-b338-4cd484b155f6","Type":"ContainerDied","Data":"9bd820698bae00820e1521389d7e85c9f91579903f903da17b0bbb829e24c442"} Oct 02 07:32:24 crc kubenswrapper[4960]: I1002 07:32:24.606474 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8c7a2f9-57bf-4520-b078-ce75a2e051ad","Type":"ContainerStarted","Data":"9bd900f71f7c49f9c12c414640951d786bf29f7c016013398508dac1dd36e0e2"} Oct 02 07:32:24 crc kubenswrapper[4960]: I1002 07:32:24.606716 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:24 crc kubenswrapper[4960]: I1002 07:32:24.690489 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=52.538093554 podStartE2EDuration="1m0.690459108s" podCreationTimestamp="2025-10-02 07:31:24 +0000 UTC" firstStartedPulling="2025-10-02 07:31:40.091434417 +0000 UTC m=+921.123380694" lastFinishedPulling="2025-10-02 07:31:48.243799951 +0000 UTC m=+929.275746248" observedRunningTime="2025-10-02 07:32:24.67702247 +0000 UTC m=+965.708968757" watchObservedRunningTime="2025-10-02 07:32:24.690459108 +0000 UTC m=+965.722405395" Oct 02 07:32:25 crc kubenswrapper[4960]: I1002 07:32:25.617960 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"36c68533-435a-4bb7-b338-4cd484b155f6","Type":"ContainerStarted","Data":"af99fe471b99be3768d7cf9a79134ba1b0f4bcfc6f1b4ed446a107e137e07773"} Oct 02 07:32:25 crc kubenswrapper[4960]: I1002 07:32:25.651782 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.921134862 podStartE2EDuration="1m0.651758302s" podCreationTimestamp="2025-10-02 07:31:25 +0000 UTC" firstStartedPulling="2025-10-02 07:31:40.569015277 +0000 UTC m=+921.600961564" lastFinishedPulling="2025-10-02 07:31:48.299638707 +0000 UTC m=+929.331585004" observedRunningTime="2025-10-02 07:32:25.649335076 +0000 UTC m=+966.681281363" watchObservedRunningTime="2025-10-02 07:32:25.651758302 +0000 UTC m=+966.683704589" Oct 02 07:32:25 crc kubenswrapper[4960]: I1002 07:32:25.975496 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-krnzn" podUID="ad314560-608d-471a-a6f6-659b9c0755d6" containerName="ovn-controller" probeResult="failure" output=< Oct 02 07:32:25 crc kubenswrapper[4960]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 07:32:25 crc kubenswrapper[4960]: > Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.087695 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.098379 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-f9tj4" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.372896 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-krnzn-config-65k92"] Oct 02 07:32:26 crc kubenswrapper[4960]: E1002 07:32:26.378074 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33dff64-7c3f-404c-9eb4-5793c6bec19b" containerName="mariadb-account-create" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.378105 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33dff64-7c3f-404c-9eb4-5793c6bec19b" containerName="mariadb-account-create" Oct 02 07:32:26 crc kubenswrapper[4960]: E1002 07:32:26.378131 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb0af53-ac0a-43d0-a651-49f1970f13f7" containerName="mariadb-account-create" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.378139 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb0af53-ac0a-43d0-a651-49f1970f13f7" containerName="mariadb-account-create" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.378365 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb0af53-ac0a-43d0-a651-49f1970f13f7" containerName="mariadb-account-create" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.378383 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d33dff64-7c3f-404c-9eb4-5793c6bec19b" containerName="mariadb-account-create" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.381731 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.388257 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.405284 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krnzn-config-65k92"] Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.481645 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-log-ovn\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.481704 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-additional-scripts\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.481820 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run-ovn\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.481856 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.481886 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqtcb\" (UniqueName: \"kubernetes.io/projected/0d382c35-4df0-4431-9823-c4e6ea261213-kube-api-access-wqtcb\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.481961 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-scripts\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.538407 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.584127 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run-ovn\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.584209 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.584253 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqtcb\" (UniqueName: \"kubernetes.io/projected/0d382c35-4df0-4431-9823-c4e6ea261213-kube-api-access-wqtcb\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.584312 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-scripts\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.584361 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-log-ovn\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.584388 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-additional-scripts\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.585220 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-additional-scripts\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.585564 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run-ovn\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.585613 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.587641 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-scripts\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.587722 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-log-ovn\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.615032 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqtcb\" (UniqueName: \"kubernetes.io/projected/0d382c35-4df0-4431-9823-c4e6ea261213-kube-api-access-wqtcb\") pod \"ovn-controller-krnzn-config-65k92\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:26 crc kubenswrapper[4960]: I1002 07:32:26.708549 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:27 crc kubenswrapper[4960]: I1002 07:32:27.188562 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krnzn-config-65k92"] Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.152221 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.152760 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.152830 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.153619 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fc62ae8f7b8e010f4106ce66a46ce004d1ca089113eca9aa199c28f45628180"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.153685 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://5fc62ae8f7b8e010f4106ce66a46ce004d1ca089113eca9aa199c28f45628180" gracePeriod=600 Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.680233 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"5fc62ae8f7b8e010f4106ce66a46ce004d1ca089113eca9aa199c28f45628180"} Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.680324 4960 scope.go:117] "RemoveContainer" containerID="856ec690ae49b3ae3dbe5a23fd799e7b1fd275afa9ab067d75708fa279072617" Oct 02 07:32:29 crc kubenswrapper[4960]: I1002 07:32:29.680161 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="5fc62ae8f7b8e010f4106ce66a46ce004d1ca089113eca9aa199c28f45628180" exitCode=0 Oct 02 07:32:30 crc kubenswrapper[4960]: I1002 07:32:30.989282 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-krnzn" podUID="ad314560-608d-471a-a6f6-659b9c0755d6" containerName="ovn-controller" probeResult="failure" output=< Oct 02 07:32:30 crc kubenswrapper[4960]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 07:32:30 crc kubenswrapper[4960]: > Oct 02 07:32:35 crc kubenswrapper[4960]: I1002 07:32:35.796930 4960 generic.go:334] "Generic (PLEG): container finished" podID="0d382c35-4df0-4431-9823-c4e6ea261213" containerID="b4f74894f81174a81e36d60ae244835da47f87bc33c704fe9b5a9a4d469e71c3" exitCode=0 Oct 02 07:32:35 crc kubenswrapper[4960]: I1002 07:32:35.797072 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn-config-65k92" event={"ID":"0d382c35-4df0-4431-9823-c4e6ea261213","Type":"ContainerDied","Data":"b4f74894f81174a81e36d60ae244835da47f87bc33c704fe9b5a9a4d469e71c3"} Oct 02 07:32:35 crc kubenswrapper[4960]: I1002 07:32:35.797843 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn-config-65k92" event={"ID":"0d382c35-4df0-4431-9823-c4e6ea261213","Type":"ContainerStarted","Data":"9c5b8d457121062030c39c3ad73414bf055b378411e1d13afd2e9edeaba8b509"} Oct 02 07:32:35 crc kubenswrapper[4960]: I1002 07:32:35.799858 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-w875j" event={"ID":"85344e88-9937-4cb2-903a-d62a423ce703","Type":"ContainerStarted","Data":"d616437001bfc596fdebbbdcf3fbbfff3a2b4e4e733592ed13c3e3511b26d010"} Oct 02 07:32:35 crc kubenswrapper[4960]: I1002 07:32:35.805696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"ae97637d73736b75a47ea930cb3cf4b53e5a377c7969711a5fc9d3103045505c"} Oct 02 07:32:35 crc kubenswrapper[4960]: I1002 07:32:35.865501 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-w875j" podStartSLOduration=1.969825586 podStartE2EDuration="15.865481704s" podCreationTimestamp="2025-10-02 07:32:20 +0000 UTC" firstStartedPulling="2025-10-02 07:32:21.105343403 +0000 UTC m=+962.137289690" lastFinishedPulling="2025-10-02 07:32:35.000999521 +0000 UTC m=+976.032945808" observedRunningTime="2025-10-02 07:32:35.859406089 +0000 UTC m=+976.891352376" watchObservedRunningTime="2025-10-02 07:32:35.865481704 +0000 UTC m=+976.897427991" Oct 02 07:32:35 crc kubenswrapper[4960]: I1002 07:32:35.974443 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-krnzn" Oct 02 07:32:36 crc kubenswrapper[4960]: I1002 07:32:36.210242 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:36 crc kubenswrapper[4960]: I1002 07:32:36.542262 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.117228 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254357 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run\") pod \"0d382c35-4df0-4431-9823-c4e6ea261213\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254471 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run-ovn\") pod \"0d382c35-4df0-4431-9823-c4e6ea261213\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254479 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run" (OuterVolumeSpecName: "var-run") pod "0d382c35-4df0-4431-9823-c4e6ea261213" (UID: "0d382c35-4df0-4431-9823-c4e6ea261213"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254500 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqtcb\" (UniqueName: \"kubernetes.io/projected/0d382c35-4df0-4431-9823-c4e6ea261213-kube-api-access-wqtcb\") pod \"0d382c35-4df0-4431-9823-c4e6ea261213\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254690 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-log-ovn\") pod \"0d382c35-4df0-4431-9823-c4e6ea261213\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254746 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-scripts\") pod \"0d382c35-4df0-4431-9823-c4e6ea261213\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254813 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-additional-scripts\") pod \"0d382c35-4df0-4431-9823-c4e6ea261213\" (UID: \"0d382c35-4df0-4431-9823-c4e6ea261213\") " Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254851 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0d382c35-4df0-4431-9823-c4e6ea261213" (UID: "0d382c35-4df0-4431-9823-c4e6ea261213"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.254882 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0d382c35-4df0-4431-9823-c4e6ea261213" (UID: "0d382c35-4df0-4431-9823-c4e6ea261213"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.255627 4960 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.255650 4960 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.255664 4960 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d382c35-4df0-4431-9823-c4e6ea261213-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.255759 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0d382c35-4df0-4431-9823-c4e6ea261213" (UID: "0d382c35-4df0-4431-9823-c4e6ea261213"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.255786 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-scripts" (OuterVolumeSpecName: "scripts") pod "0d382c35-4df0-4431-9823-c4e6ea261213" (UID: "0d382c35-4df0-4431-9823-c4e6ea261213"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.260924 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d382c35-4df0-4431-9823-c4e6ea261213-kube-api-access-wqtcb" (OuterVolumeSpecName: "kube-api-access-wqtcb") pod "0d382c35-4df0-4431-9823-c4e6ea261213" (UID: "0d382c35-4df0-4431-9823-c4e6ea261213"). InnerVolumeSpecName "kube-api-access-wqtcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.358044 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqtcb\" (UniqueName: \"kubernetes.io/projected/0d382c35-4df0-4431-9823-c4e6ea261213-kube-api-access-wqtcb\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.358091 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.358113 4960 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d382c35-4df0-4431-9823-c4e6ea261213-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.824332 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn-config-65k92" event={"ID":"0d382c35-4df0-4431-9823-c4e6ea261213","Type":"ContainerDied","Data":"9c5b8d457121062030c39c3ad73414bf055b378411e1d13afd2e9edeaba8b509"} Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.824867 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c5b8d457121062030c39c3ad73414bf055b378411e1d13afd2e9edeaba8b509" Oct 02 07:32:37 crc kubenswrapper[4960]: I1002 07:32:37.824965 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-65k92" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.226603 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-krnzn-config-65k92"] Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.233186 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-krnzn-config-65k92"] Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.340761 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d382c35-4df0-4431-9823-c4e6ea261213" path="/var/lib/kubelet/pods/0d382c35-4df0-4431-9823-c4e6ea261213/volumes" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.360165 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-krnzn-config-kwvgd"] Oct 02 07:32:38 crc kubenswrapper[4960]: E1002 07:32:38.360600 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d382c35-4df0-4431-9823-c4e6ea261213" containerName="ovn-config" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.360620 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d382c35-4df0-4431-9823-c4e6ea261213" containerName="ovn-config" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.360798 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d382c35-4df0-4431-9823-c4e6ea261213" containerName="ovn-config" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.361414 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.363388 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.374383 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-log-ovn\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.374444 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-additional-scripts\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.374474 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run-ovn\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.374502 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.374541 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc447\" (UniqueName: \"kubernetes.io/projected/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-kube-api-access-qc447\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.374579 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-scripts\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.377271 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krnzn-config-kwvgd"] Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.476710 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-additional-scripts\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.476761 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.476789 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run-ovn\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.476826 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc447\" (UniqueName: \"kubernetes.io/projected/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-kube-api-access-qc447\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.476864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-scripts\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.476947 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-log-ovn\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.477314 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-log-ovn\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.477317 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.477370 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run-ovn\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.477844 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-additional-scripts\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.479266 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-scripts\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.499616 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc447\" (UniqueName: \"kubernetes.io/projected/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-kube-api-access-qc447\") pod \"ovn-controller-krnzn-config-kwvgd\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.680732 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.778765 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-2kx2k"] Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.780355 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2kx2k" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.796476 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2kx2k"] Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.884099 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnb88\" (UniqueName: \"kubernetes.io/projected/ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15-kube-api-access-dnb88\") pod \"cinder-db-create-2kx2k\" (UID: \"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15\") " pod="openstack/cinder-db-create-2kx2k" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.884479 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-lnmn2"] Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.885948 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnmn2" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.907143 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lnmn2"] Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.986102 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnb88\" (UniqueName: \"kubernetes.io/projected/ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15-kube-api-access-dnb88\") pod \"cinder-db-create-2kx2k\" (UID: \"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15\") " pod="openstack/cinder-db-create-2kx2k" Oct 02 07:32:38 crc kubenswrapper[4960]: I1002 07:32:38.986171 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n8pg\" (UniqueName: \"kubernetes.io/projected/7df3811d-f779-4df9-a3cc-6a74d7dc00b3-kube-api-access-5n8pg\") pod \"barbican-db-create-lnmn2\" (UID: \"7df3811d-f779-4df9-a3cc-6a74d7dc00b3\") " pod="openstack/barbican-db-create-lnmn2" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.016725 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnb88\" (UniqueName: \"kubernetes.io/projected/ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15-kube-api-access-dnb88\") pod \"cinder-db-create-2kx2k\" (UID: \"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15\") " pod="openstack/cinder-db-create-2kx2k" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.085572 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krnzn-config-kwvgd"] Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.087427 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n8pg\" (UniqueName: \"kubernetes.io/projected/7df3811d-f779-4df9-a3cc-6a74d7dc00b3-kube-api-access-5n8pg\") pod \"barbican-db-create-lnmn2\" (UID: \"7df3811d-f779-4df9-a3cc-6a74d7dc00b3\") " pod="openstack/barbican-db-create-lnmn2" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.102894 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-mfsh5"] Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.104296 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfsh5" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.105185 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2kx2k" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.111550 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mfsh5"] Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.112704 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n8pg\" (UniqueName: \"kubernetes.io/projected/7df3811d-f779-4df9-a3cc-6a74d7dc00b3-kube-api-access-5n8pg\") pod \"barbican-db-create-lnmn2\" (UID: \"7df3811d-f779-4df9-a3cc-6a74d7dc00b3\") " pod="openstack/barbican-db-create-lnmn2" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.169344 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bcvcn"] Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.170760 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.178754 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.178967 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.179076 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mplxm" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.179373 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.186806 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bcvcn"] Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.232274 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnmn2" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.290739 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-config-data\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.290845 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfrhl\" (UniqueName: \"kubernetes.io/projected/f0aea82a-faa5-46de-a545-54b2ac12c1bd-kube-api-access-xfrhl\") pod \"neutron-db-create-mfsh5\" (UID: \"f0aea82a-faa5-46de-a545-54b2ac12c1bd\") " pod="openstack/neutron-db-create-mfsh5" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.290892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-combined-ca-bundle\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.290957 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzrpl\" (UniqueName: \"kubernetes.io/projected/39377f09-4342-42bc-87a5-17639bdd8f99-kube-api-access-pzrpl\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.399906 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-combined-ca-bundle\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.400797 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzrpl\" (UniqueName: \"kubernetes.io/projected/39377f09-4342-42bc-87a5-17639bdd8f99-kube-api-access-pzrpl\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.400842 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-config-data\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.400890 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfrhl\" (UniqueName: \"kubernetes.io/projected/f0aea82a-faa5-46de-a545-54b2ac12c1bd-kube-api-access-xfrhl\") pod \"neutron-db-create-mfsh5\" (UID: \"f0aea82a-faa5-46de-a545-54b2ac12c1bd\") " pod="openstack/neutron-db-create-mfsh5" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.412886 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-config-data\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.419854 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-combined-ca-bundle\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.429147 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzrpl\" (UniqueName: \"kubernetes.io/projected/39377f09-4342-42bc-87a5-17639bdd8f99-kube-api-access-pzrpl\") pod \"keystone-db-sync-bcvcn\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.438750 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfrhl\" (UniqueName: \"kubernetes.io/projected/f0aea82a-faa5-46de-a545-54b2ac12c1bd-kube-api-access-xfrhl\") pod \"neutron-db-create-mfsh5\" (UID: \"f0aea82a-faa5-46de-a545-54b2ac12c1bd\") " pod="openstack/neutron-db-create-mfsh5" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.440524 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfsh5" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.491464 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.699755 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2kx2k"] Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.851628 4960 generic.go:334] "Generic (PLEG): container finished" podID="bc8fcd1a-0470-4752-8ff4-e9edaab1711d" containerID="628286f25086ad1a428009e59f9e7361937a92068d53660e27afe48bf4ad4032" exitCode=0 Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.852273 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn-config-kwvgd" event={"ID":"bc8fcd1a-0470-4752-8ff4-e9edaab1711d","Type":"ContainerDied","Data":"628286f25086ad1a428009e59f9e7361937a92068d53660e27afe48bf4ad4032"} Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.852313 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn-config-kwvgd" event={"ID":"bc8fcd1a-0470-4752-8ff4-e9edaab1711d","Type":"ContainerStarted","Data":"4e9a10ab05f2dcaf9077225b04fd2d0aa76eda4aa3f2e69962102857dd362e33"} Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.853912 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2kx2k" event={"ID":"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15","Type":"ContainerStarted","Data":"c989f90978b32dd8048fa9e3acc4fb909ac65faf4a6fca6d54bf1951d51ed8e9"} Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.856768 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lnmn2"] Oct 02 07:32:39 crc kubenswrapper[4960]: W1002 07:32:39.894484 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7df3811d_f779_4df9_a3cc_6a74d7dc00b3.slice/crio-34268375508d66aecef79a34f7dea9fb9e220d9be7e30415f9db805c58d38daf WatchSource:0}: Error finding container 34268375508d66aecef79a34f7dea9fb9e220d9be7e30415f9db805c58d38daf: Status 404 returned error can't find the container with id 34268375508d66aecef79a34f7dea9fb9e220d9be7e30415f9db805c58d38daf Oct 02 07:32:39 crc kubenswrapper[4960]: W1002 07:32:39.949812 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0aea82a_faa5_46de_a545_54b2ac12c1bd.slice/crio-c98a717751daa6968700a4198554302f506a6829e2ab3339eb47ab43a95c4ab8 WatchSource:0}: Error finding container c98a717751daa6968700a4198554302f506a6829e2ab3339eb47ab43a95c4ab8: Status 404 returned error can't find the container with id c98a717751daa6968700a4198554302f506a6829e2ab3339eb47ab43a95c4ab8 Oct 02 07:32:39 crc kubenswrapper[4960]: I1002 07:32:39.951080 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mfsh5"] Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.050907 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bcvcn"] Oct 02 07:32:40 crc kubenswrapper[4960]: W1002 07:32:40.058465 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39377f09_4342_42bc_87a5_17639bdd8f99.slice/crio-fe5e62e9894e878e0820959e0e6a268f4c33511a447e834c3866f618d1348ac8 WatchSource:0}: Error finding container fe5e62e9894e878e0820959e0e6a268f4c33511a447e834c3866f618d1348ac8: Status 404 returned error can't find the container with id fe5e62e9894e878e0820959e0e6a268f4c33511a447e834c3866f618d1348ac8 Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.865132 4960 generic.go:334] "Generic (PLEG): container finished" podID="f0aea82a-faa5-46de-a545-54b2ac12c1bd" containerID="7bf1bbac7d2f8d1ba179c0d7907a146854bc10ec869b358abb0682ab64ea5403" exitCode=0 Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.865360 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mfsh5" event={"ID":"f0aea82a-faa5-46de-a545-54b2ac12c1bd","Type":"ContainerDied","Data":"7bf1bbac7d2f8d1ba179c0d7907a146854bc10ec869b358abb0682ab64ea5403"} Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.865472 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mfsh5" event={"ID":"f0aea82a-faa5-46de-a545-54b2ac12c1bd","Type":"ContainerStarted","Data":"c98a717751daa6968700a4198554302f506a6829e2ab3339eb47ab43a95c4ab8"} Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.868044 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bcvcn" event={"ID":"39377f09-4342-42bc-87a5-17639bdd8f99","Type":"ContainerStarted","Data":"fe5e62e9894e878e0820959e0e6a268f4c33511a447e834c3866f618d1348ac8"} Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.871626 4960 generic.go:334] "Generic (PLEG): container finished" podID="ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15" containerID="1501122a8ea6e8e195ed6a8df5fa7e8886e706c2b22fe0f37b5daacd592784c6" exitCode=0 Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.871669 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2kx2k" event={"ID":"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15","Type":"ContainerDied","Data":"1501122a8ea6e8e195ed6a8df5fa7e8886e706c2b22fe0f37b5daacd592784c6"} Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.873196 4960 generic.go:334] "Generic (PLEG): container finished" podID="7df3811d-f779-4df9-a3cc-6a74d7dc00b3" containerID="bfde7721b34d159d48b79949ece39ce5e626bb613a808a79f0598c73999c8955" exitCode=0 Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.873386 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lnmn2" event={"ID":"7df3811d-f779-4df9-a3cc-6a74d7dc00b3","Type":"ContainerDied","Data":"bfde7721b34d159d48b79949ece39ce5e626bb613a808a79f0598c73999c8955"} Oct 02 07:32:40 crc kubenswrapper[4960]: I1002 07:32:40.873411 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lnmn2" event={"ID":"7df3811d-f779-4df9-a3cc-6a74d7dc00b3","Type":"ContainerStarted","Data":"34268375508d66aecef79a34f7dea9fb9e220d9be7e30415f9db805c58d38daf"} Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.226122 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.238558 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run\") pod \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.238715 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-additional-scripts\") pod \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.238773 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-scripts\") pod \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.238799 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc447\" (UniqueName: \"kubernetes.io/projected/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-kube-api-access-qc447\") pod \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.238816 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run-ovn\") pod \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.238889 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-log-ovn\") pod \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\" (UID: \"bc8fcd1a-0470-4752-8ff4-e9edaab1711d\") " Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.239306 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "bc8fcd1a-0470-4752-8ff4-e9edaab1711d" (UID: "bc8fcd1a-0470-4752-8ff4-e9edaab1711d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.239342 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run" (OuterVolumeSpecName: "var-run") pod "bc8fcd1a-0470-4752-8ff4-e9edaab1711d" (UID: "bc8fcd1a-0470-4752-8ff4-e9edaab1711d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.240233 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "bc8fcd1a-0470-4752-8ff4-e9edaab1711d" (UID: "bc8fcd1a-0470-4752-8ff4-e9edaab1711d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.240889 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-scripts" (OuterVolumeSpecName: "scripts") pod "bc8fcd1a-0470-4752-8ff4-e9edaab1711d" (UID: "bc8fcd1a-0470-4752-8ff4-e9edaab1711d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.242505 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "bc8fcd1a-0470-4752-8ff4-e9edaab1711d" (UID: "bc8fcd1a-0470-4752-8ff4-e9edaab1711d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.250606 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-kube-api-access-qc447" (OuterVolumeSpecName: "kube-api-access-qc447") pod "bc8fcd1a-0470-4752-8ff4-e9edaab1711d" (UID: "bc8fcd1a-0470-4752-8ff4-e9edaab1711d"). InnerVolumeSpecName "kube-api-access-qc447". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.342238 4960 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.342279 4960 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.342296 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.342311 4960 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.342326 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc447\" (UniqueName: \"kubernetes.io/projected/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-kube-api-access-qc447\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.342336 4960 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fcd1a-0470-4752-8ff4-e9edaab1711d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.887002 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krnzn-config-kwvgd" Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.887165 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krnzn-config-kwvgd" event={"ID":"bc8fcd1a-0470-4752-8ff4-e9edaab1711d","Type":"ContainerDied","Data":"4e9a10ab05f2dcaf9077225b04fd2d0aa76eda4aa3f2e69962102857dd362e33"} Oct 02 07:32:41 crc kubenswrapper[4960]: I1002 07:32:41.889996 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e9a10ab05f2dcaf9077225b04fd2d0aa76eda4aa3f2e69962102857dd362e33" Oct 02 07:32:42 crc kubenswrapper[4960]: I1002 07:32:42.349653 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-krnzn-config-kwvgd"] Oct 02 07:32:42 crc kubenswrapper[4960]: I1002 07:32:42.355867 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-krnzn-config-kwvgd"] Oct 02 07:32:42 crc kubenswrapper[4960]: I1002 07:32:42.901432 4960 generic.go:334] "Generic (PLEG): container finished" podID="85344e88-9937-4cb2-903a-d62a423ce703" containerID="d616437001bfc596fdebbbdcf3fbbfff3a2b4e4e733592ed13c3e3511b26d010" exitCode=0 Oct 02 07:32:42 crc kubenswrapper[4960]: I1002 07:32:42.901530 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-w875j" event={"ID":"85344e88-9937-4cb2-903a-d62a423ce703","Type":"ContainerDied","Data":"d616437001bfc596fdebbbdcf3fbbfff3a2b4e4e733592ed13c3e3511b26d010"} Oct 02 07:32:44 crc kubenswrapper[4960]: I1002 07:32:44.350564 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc8fcd1a-0470-4752-8ff4-e9edaab1711d" path="/var/lib/kubelet/pods/bc8fcd1a-0470-4752-8ff4-e9edaab1711d/volumes" Oct 02 07:32:44 crc kubenswrapper[4960]: I1002 07:32:44.924749 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2kx2k" event={"ID":"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15","Type":"ContainerDied","Data":"c989f90978b32dd8048fa9e3acc4fb909ac65faf4a6fca6d54bf1951d51ed8e9"} Oct 02 07:32:44 crc kubenswrapper[4960]: I1002 07:32:44.925292 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c989f90978b32dd8048fa9e3acc4fb909ac65faf4a6fca6d54bf1951d51ed8e9" Oct 02 07:32:44 crc kubenswrapper[4960]: I1002 07:32:44.927897 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lnmn2" event={"ID":"7df3811d-f779-4df9-a3cc-6a74d7dc00b3","Type":"ContainerDied","Data":"34268375508d66aecef79a34f7dea9fb9e220d9be7e30415f9db805c58d38daf"} Oct 02 07:32:44 crc kubenswrapper[4960]: I1002 07:32:44.927934 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34268375508d66aecef79a34f7dea9fb9e220d9be7e30415f9db805c58d38daf" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.015912 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2kx2k" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.021878 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnmn2" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.043222 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfsh5" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.124888 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnb88\" (UniqueName: \"kubernetes.io/projected/ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15-kube-api-access-dnb88\") pod \"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15\" (UID: \"ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15\") " Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.124954 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n8pg\" (UniqueName: \"kubernetes.io/projected/7df3811d-f779-4df9-a3cc-6a74d7dc00b3-kube-api-access-5n8pg\") pod \"7df3811d-f779-4df9-a3cc-6a74d7dc00b3\" (UID: \"7df3811d-f779-4df9-a3cc-6a74d7dc00b3\") " Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.125181 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfrhl\" (UniqueName: \"kubernetes.io/projected/f0aea82a-faa5-46de-a545-54b2ac12c1bd-kube-api-access-xfrhl\") pod \"f0aea82a-faa5-46de-a545-54b2ac12c1bd\" (UID: \"f0aea82a-faa5-46de-a545-54b2ac12c1bd\") " Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.132130 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-w875j" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.132470 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df3811d-f779-4df9-a3cc-6a74d7dc00b3-kube-api-access-5n8pg" (OuterVolumeSpecName: "kube-api-access-5n8pg") pod "7df3811d-f779-4df9-a3cc-6a74d7dc00b3" (UID: "7df3811d-f779-4df9-a3cc-6a74d7dc00b3"). InnerVolumeSpecName "kube-api-access-5n8pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.132627 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15-kube-api-access-dnb88" (OuterVolumeSpecName: "kube-api-access-dnb88") pod "ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15" (UID: "ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15"). InnerVolumeSpecName "kube-api-access-dnb88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.133907 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0aea82a-faa5-46de-a545-54b2ac12c1bd-kube-api-access-xfrhl" (OuterVolumeSpecName: "kube-api-access-xfrhl") pod "f0aea82a-faa5-46de-a545-54b2ac12c1bd" (UID: "f0aea82a-faa5-46de-a545-54b2ac12c1bd"). InnerVolumeSpecName "kube-api-access-xfrhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.226430 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-787s4\" (UniqueName: \"kubernetes.io/projected/85344e88-9937-4cb2-903a-d62a423ce703-kube-api-access-787s4\") pod \"85344e88-9937-4cb2-903a-d62a423ce703\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.226939 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-config-data\") pod \"85344e88-9937-4cb2-903a-d62a423ce703\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.227004 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-combined-ca-bundle\") pod \"85344e88-9937-4cb2-903a-d62a423ce703\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.227069 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-db-sync-config-data\") pod \"85344e88-9937-4cb2-903a-d62a423ce703\" (UID: \"85344e88-9937-4cb2-903a-d62a423ce703\") " Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.227405 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfrhl\" (UniqueName: \"kubernetes.io/projected/f0aea82a-faa5-46de-a545-54b2ac12c1bd-kube-api-access-xfrhl\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.227419 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnb88\" (UniqueName: \"kubernetes.io/projected/ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15-kube-api-access-dnb88\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.227433 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n8pg\" (UniqueName: \"kubernetes.io/projected/7df3811d-f779-4df9-a3cc-6a74d7dc00b3-kube-api-access-5n8pg\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.230486 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "85344e88-9937-4cb2-903a-d62a423ce703" (UID: "85344e88-9937-4cb2-903a-d62a423ce703"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.232338 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85344e88-9937-4cb2-903a-d62a423ce703-kube-api-access-787s4" (OuterVolumeSpecName: "kube-api-access-787s4") pod "85344e88-9937-4cb2-903a-d62a423ce703" (UID: "85344e88-9937-4cb2-903a-d62a423ce703"). InnerVolumeSpecName "kube-api-access-787s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.297244 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85344e88-9937-4cb2-903a-d62a423ce703" (UID: "85344e88-9937-4cb2-903a-d62a423ce703"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.323721 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-config-data" (OuterVolumeSpecName: "config-data") pod "85344e88-9937-4cb2-903a-d62a423ce703" (UID: "85344e88-9937-4cb2-903a-d62a423ce703"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.329168 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-787s4\" (UniqueName: \"kubernetes.io/projected/85344e88-9937-4cb2-903a-d62a423ce703-kube-api-access-787s4\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.329220 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.329237 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.329252 4960 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85344e88-9937-4cb2-903a-d62a423ce703-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.939773 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mfsh5" event={"ID":"f0aea82a-faa5-46de-a545-54b2ac12c1bd","Type":"ContainerDied","Data":"c98a717751daa6968700a4198554302f506a6829e2ab3339eb47ab43a95c4ab8"} Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.940156 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c98a717751daa6968700a4198554302f506a6829e2ab3339eb47ab43a95c4ab8" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.939847 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfsh5" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.944138 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bcvcn" event={"ID":"39377f09-4342-42bc-87a5-17639bdd8f99","Type":"ContainerStarted","Data":"b1c20644b9ca5b6b0b07d4df04a174a2b4052b9fe8cc4564c0466c18c894e1f5"} Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.946402 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-w875j" event={"ID":"85344e88-9937-4cb2-903a-d62a423ce703","Type":"ContainerDied","Data":"28fb8f27a341fae9a04752e8fe961caef579eb6f43f0d9255e68a888e062e7fa"} Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.946533 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28fb8f27a341fae9a04752e8fe961caef579eb6f43f0d9255e68a888e062e7fa" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.946599 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-w875j" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.946537 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2kx2k" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.946473 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnmn2" Oct 02 07:32:45 crc kubenswrapper[4960]: I1002 07:32:45.987881 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bcvcn" podStartSLOduration=1.887982051 podStartE2EDuration="6.987862824s" podCreationTimestamp="2025-10-02 07:32:39 +0000 UTC" firstStartedPulling="2025-10-02 07:32:40.062331609 +0000 UTC m=+981.094277896" lastFinishedPulling="2025-10-02 07:32:45.162212362 +0000 UTC m=+986.194158669" observedRunningTime="2025-10-02 07:32:45.973284793 +0000 UTC m=+987.005231080" watchObservedRunningTime="2025-10-02 07:32:45.987862824 +0000 UTC m=+987.019809111" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.602755 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9dd498fbf-fgq6l"] Oct 02 07:32:46 crc kubenswrapper[4960]: E1002 07:32:46.603177 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df3811d-f779-4df9-a3cc-6a74d7dc00b3" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603196 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df3811d-f779-4df9-a3cc-6a74d7dc00b3" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: E1002 07:32:46.603207 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc8fcd1a-0470-4752-8ff4-e9edaab1711d" containerName="ovn-config" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603215 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc8fcd1a-0470-4752-8ff4-e9edaab1711d" containerName="ovn-config" Oct 02 07:32:46 crc kubenswrapper[4960]: E1002 07:32:46.603239 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0aea82a-faa5-46de-a545-54b2ac12c1bd" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603246 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0aea82a-faa5-46de-a545-54b2ac12c1bd" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: E1002 07:32:46.603265 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85344e88-9937-4cb2-903a-d62a423ce703" containerName="glance-db-sync" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603271 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="85344e88-9937-4cb2-903a-d62a423ce703" containerName="glance-db-sync" Oct 02 07:32:46 crc kubenswrapper[4960]: E1002 07:32:46.603281 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603289 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603454 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0aea82a-faa5-46de-a545-54b2ac12c1bd" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603474 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df3811d-f779-4df9-a3cc-6a74d7dc00b3" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603484 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15" containerName="mariadb-database-create" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603496 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="85344e88-9937-4cb2-903a-d62a423ce703" containerName="glance-db-sync" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.603505 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc8fcd1a-0470-4752-8ff4-e9edaab1711d" containerName="ovn-config" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.604631 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.635597 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9dd498fbf-fgq6l"] Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.653505 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-sb\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.653561 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-dns-svc\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.653604 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj72g\" (UniqueName: \"kubernetes.io/projected/d58a2b49-97bb-4175-8b96-5f49fa95848b-kube-api-access-tj72g\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.653643 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-config\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.653706 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-nb\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.755489 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-nb\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.755599 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-sb\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.755645 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-dns-svc\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.755693 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj72g\" (UniqueName: \"kubernetes.io/projected/d58a2b49-97bb-4175-8b96-5f49fa95848b-kube-api-access-tj72g\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.755746 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-config\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.756879 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-nb\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.757021 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-dns-svc\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.757257 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-sb\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.757340 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-config\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.776837 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj72g\" (UniqueName: \"kubernetes.io/projected/d58a2b49-97bb-4175-8b96-5f49fa95848b-kube-api-access-tj72g\") pod \"dnsmasq-dns-9dd498fbf-fgq6l\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:46 crc kubenswrapper[4960]: I1002 07:32:46.978036 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:47 crc kubenswrapper[4960]: I1002 07:32:47.625682 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9dd498fbf-fgq6l"] Oct 02 07:32:47 crc kubenswrapper[4960]: W1002 07:32:47.632846 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd58a2b49_97bb_4175_8b96_5f49fa95848b.slice/crio-78e1aecd8165e0ad141bedac2154abf83aa125e769acf0e0dd93d3652d3bdf9a WatchSource:0}: Error finding container 78e1aecd8165e0ad141bedac2154abf83aa125e769acf0e0dd93d3652d3bdf9a: Status 404 returned error can't find the container with id 78e1aecd8165e0ad141bedac2154abf83aa125e769acf0e0dd93d3652d3bdf9a Oct 02 07:32:47 crc kubenswrapper[4960]: I1002 07:32:47.974314 4960 generic.go:334] "Generic (PLEG): container finished" podID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerID="941e3986ce90567e2192200768c0e7bcd3ab1a6a6c75ad93f1e48dceb1d275ed" exitCode=0 Oct 02 07:32:47 crc kubenswrapper[4960]: I1002 07:32:47.974421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" event={"ID":"d58a2b49-97bb-4175-8b96-5f49fa95848b","Type":"ContainerDied","Data":"941e3986ce90567e2192200768c0e7bcd3ab1a6a6c75ad93f1e48dceb1d275ed"} Oct 02 07:32:47 crc kubenswrapper[4960]: I1002 07:32:47.974872 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" event={"ID":"d58a2b49-97bb-4175-8b96-5f49fa95848b","Type":"ContainerStarted","Data":"78e1aecd8165e0ad141bedac2154abf83aa125e769acf0e0dd93d3652d3bdf9a"} Oct 02 07:32:49 crc kubenswrapper[4960]: I1002 07:32:49.021522 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" event={"ID":"d58a2b49-97bb-4175-8b96-5f49fa95848b","Type":"ContainerStarted","Data":"e763f445778bfd42a76b46f2c787c94f68f8dd5a33332d5029008b8f59fa4f2b"} Oct 02 07:32:49 crc kubenswrapper[4960]: I1002 07:32:49.026094 4960 generic.go:334] "Generic (PLEG): container finished" podID="39377f09-4342-42bc-87a5-17639bdd8f99" containerID="b1c20644b9ca5b6b0b07d4df04a174a2b4052b9fe8cc4564c0466c18c894e1f5" exitCode=0 Oct 02 07:32:49 crc kubenswrapper[4960]: I1002 07:32:49.026157 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bcvcn" event={"ID":"39377f09-4342-42bc-87a5-17639bdd8f99","Type":"ContainerDied","Data":"b1c20644b9ca5b6b0b07d4df04a174a2b4052b9fe8cc4564c0466c18c894e1f5"} Oct 02 07:32:49 crc kubenswrapper[4960]: I1002 07:32:49.045396 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" podStartSLOduration=3.045369854 podStartE2EDuration="3.045369854s" podCreationTimestamp="2025-10-02 07:32:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:32:49.044069466 +0000 UTC m=+990.076015773" watchObservedRunningTime="2025-10-02 07:32:49.045369854 +0000 UTC m=+990.077316151" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.035215 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.406435 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.432139 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-config-data\") pod \"39377f09-4342-42bc-87a5-17639bdd8f99\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.432213 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-combined-ca-bundle\") pod \"39377f09-4342-42bc-87a5-17639bdd8f99\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.432252 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzrpl\" (UniqueName: \"kubernetes.io/projected/39377f09-4342-42bc-87a5-17639bdd8f99-kube-api-access-pzrpl\") pod \"39377f09-4342-42bc-87a5-17639bdd8f99\" (UID: \"39377f09-4342-42bc-87a5-17639bdd8f99\") " Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.439380 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39377f09-4342-42bc-87a5-17639bdd8f99-kube-api-access-pzrpl" (OuterVolumeSpecName: "kube-api-access-pzrpl") pod "39377f09-4342-42bc-87a5-17639bdd8f99" (UID: "39377f09-4342-42bc-87a5-17639bdd8f99"). InnerVolumeSpecName "kube-api-access-pzrpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.460673 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39377f09-4342-42bc-87a5-17639bdd8f99" (UID: "39377f09-4342-42bc-87a5-17639bdd8f99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.492004 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-config-data" (OuterVolumeSpecName: "config-data") pod "39377f09-4342-42bc-87a5-17639bdd8f99" (UID: "39377f09-4342-42bc-87a5-17639bdd8f99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.534866 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.535119 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzrpl\" (UniqueName: \"kubernetes.io/projected/39377f09-4342-42bc-87a5-17639bdd8f99-kube-api-access-pzrpl\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:50 crc kubenswrapper[4960]: I1002 07:32:50.535190 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39377f09-4342-42bc-87a5-17639bdd8f99-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.047964 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bcvcn" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.048212 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bcvcn" event={"ID":"39377f09-4342-42bc-87a5-17639bdd8f99","Type":"ContainerDied","Data":"fe5e62e9894e878e0820959e0e6a268f4c33511a447e834c3866f618d1348ac8"} Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.048588 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe5e62e9894e878e0820959e0e6a268f4c33511a447e834c3866f618d1348ac8" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.347765 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fj8mf"] Oct 02 07:32:51 crc kubenswrapper[4960]: E1002 07:32:51.348198 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39377f09-4342-42bc-87a5-17639bdd8f99" containerName="keystone-db-sync" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.348215 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="39377f09-4342-42bc-87a5-17639bdd8f99" containerName="keystone-db-sync" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.348485 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="39377f09-4342-42bc-87a5-17639bdd8f99" containerName="keystone-db-sync" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.349155 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.351561 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.353993 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mplxm" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.353999 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.362039 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.363713 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9dd498fbf-fgq6l"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.388189 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fj8mf"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.415154 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-747ffb4f4f-zhxrd"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.416954 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.436467 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-747ffb4f4f-zhxrd"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.456611 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-scripts\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.456691 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2958\" (UniqueName: \"kubernetes.io/projected/04c274e5-afed-4908-9f2e-1cd8165f9e6d-kube-api-access-k2958\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.456733 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-fernet-keys\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.456777 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-combined-ca-bundle\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.456822 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-credential-keys\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.456850 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-nb\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.456957 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-config\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.457031 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-sb\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.457063 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kxxj\" (UniqueName: \"kubernetes.io/projected/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-kube-api-access-7kxxj\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.457108 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-dns-svc\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.457143 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-config-data\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558230 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-scripts\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558301 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2958\" (UniqueName: \"kubernetes.io/projected/04c274e5-afed-4908-9f2e-1cd8165f9e6d-kube-api-access-k2958\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558332 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-fernet-keys\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558359 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-combined-ca-bundle\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558390 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-credential-keys\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558413 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-nb\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558459 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-config\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558489 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-sb\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558516 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kxxj\" (UniqueName: \"kubernetes.io/projected/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-kube-api-access-7kxxj\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558547 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-dns-svc\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.558574 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-config-data\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.559691 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-nb\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.559728 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-sb\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.559731 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-dns-svc\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.560126 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-config\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.571634 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-scripts\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.581161 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-fernet-keys\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.586625 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-credential-keys\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.587896 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-844dfdd985-j275s"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.591430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kxxj\" (UniqueName: \"kubernetes.io/projected/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-kube-api-access-7kxxj\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.594801 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-config-data\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.596879 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.613440 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.613935 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-p2xmn" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.614093 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.614217 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.615881 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-combined-ca-bundle\") pod \"keystone-bootstrap-fj8mf\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.626099 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-844dfdd985-j275s"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.636604 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2958\" (UniqueName: \"kubernetes.io/projected/04c274e5-afed-4908-9f2e-1cd8165f9e6d-kube-api-access-k2958\") pod \"dnsmasq-dns-747ffb4f4f-zhxrd\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.658425 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.660943 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.667983 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.669122 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-scripts\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.669162 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-logs\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.669236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-config-data\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.669306 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbczx\" (UniqueName: \"kubernetes.io/projected/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-kube-api-access-pbczx\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.669399 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-horizon-secret-key\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.670358 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.681095 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.691643 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.747263 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.787582 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-scripts\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.787793 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-scripts\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.787819 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-logs\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.787854 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-config-data\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.788990 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-config-data\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.789181 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbczx\" (UniqueName: \"kubernetes.io/projected/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-kube-api-access-pbczx\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.789320 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lxqm\" (UniqueName: \"kubernetes.io/projected/34621790-60f4-4239-a57d-941591637b6e-kube-api-access-9lxqm\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.789363 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-log-httpd\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.789390 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.789438 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-run-httpd\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.789463 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.789492 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-horizon-secret-key\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.798549 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-horizon-secret-key\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.799651 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-scripts\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.818433 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-logs\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.821266 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-config-data\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.851253 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-sl4rw"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.859734 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbczx\" (UniqueName: \"kubernetes.io/projected/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-kube-api-access-pbczx\") pod \"horizon-844dfdd985-j275s\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.860905 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.872139 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.872671 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.872611 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wd5ms" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.894315 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lxqm\" (UniqueName: \"kubernetes.io/projected/34621790-60f4-4239-a57d-941591637b6e-kube-api-access-9lxqm\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.894355 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-log-httpd\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.894379 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.894401 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-run-httpd\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.894427 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895213 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-scripts\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895243 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-combined-ca-bundle\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895264 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8spbc\" (UniqueName: \"kubernetes.io/projected/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-kube-api-access-8spbc\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895302 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-config-data\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895344 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-scripts\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895361 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-logs\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895409 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-config-data\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.895991 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-log-httpd\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.899264 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.899463 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-run-httpd\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.910188 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-sl4rw"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.910328 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-config-data\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.910931 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.914394 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-scripts\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.922438 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-747ffb4f4f-zhxrd"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.932114 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-796bc46545-gltd7"] Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.940768 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:51 crc kubenswrapper[4960]: I1002 07:32:51.969759 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-796bc46545-gltd7"] Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005041 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-config\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005144 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-combined-ca-bundle\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005187 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8spbc\" (UniqueName: \"kubernetes.io/projected/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-kube-api-access-8spbc\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005258 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-nb\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005293 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-scripts\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005314 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-logs\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005344 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-dns-svc\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005375 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-config-data\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005398 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdcqj\" (UniqueName: \"kubernetes.io/projected/0a970f1b-0b01-4b2c-821b-a2c9660e1845-kube-api-access-qdcqj\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005421 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-sb\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.005834 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lxqm\" (UniqueName: \"kubernetes.io/projected/34621790-60f4-4239-a57d-941591637b6e-kube-api-access-9lxqm\") pod \"ceilometer-0\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " pod="openstack/ceilometer-0" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.006436 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-logs\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.014257 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-scripts\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.033523 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-combined-ca-bundle\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.040233 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-config-data\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.077496 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8spbc\" (UniqueName: \"kubernetes.io/projected/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-kube-api-access-8spbc\") pod \"placement-db-sync-sl4rw\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.078729 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" podUID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerName="dnsmasq-dns" containerID="cri-o://e763f445778bfd42a76b46f2c787c94f68f8dd5a33332d5029008b8f59fa4f2b" gracePeriod=10 Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.109083 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-745f47d489-qq69f"] Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.111375 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.127082 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-sb\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.127168 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-config\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.127228 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-nb\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.127258 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-dns-svc\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.127287 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdcqj\" (UniqueName: \"kubernetes.io/projected/0a970f1b-0b01-4b2c-821b-a2c9660e1845-kube-api-access-qdcqj\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.156323 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.183676 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-nb\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.184182 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.186356 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdcqj\" (UniqueName: \"kubernetes.io/projected/0a970f1b-0b01-4b2c-821b-a2c9660e1845-kube-api-access-qdcqj\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.186446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-config\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.187079 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-sb\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.194117 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sl4rw" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.197069 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-dns-svc\") pod \"dnsmasq-dns-796bc46545-gltd7\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.197148 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-745f47d489-qq69f"] Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.229245 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ac3e0e9-5501-4af6-a745-3ef391fcac43-horizon-secret-key\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.229318 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25qmf\" (UniqueName: \"kubernetes.io/projected/6ac3e0e9-5501-4af6-a745-3ef391fcac43-kube-api-access-25qmf\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.229368 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-config-data\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.229416 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-scripts\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.229464 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac3e0e9-5501-4af6-a745-3ef391fcac43-logs\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.333300 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-config-data\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.333720 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-scripts\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.333853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac3e0e9-5501-4af6-a745-3ef391fcac43-logs\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.333969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ac3e0e9-5501-4af6-a745-3ef391fcac43-horizon-secret-key\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.334474 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25qmf\" (UniqueName: \"kubernetes.io/projected/6ac3e0e9-5501-4af6-a745-3ef391fcac43-kube-api-access-25qmf\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.335180 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-scripts\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.336188 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-config-data\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.340400 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac3e0e9-5501-4af6-a745-3ef391fcac43-logs\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.349118 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.361574 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ac3e0e9-5501-4af6-a745-3ef391fcac43-horizon-secret-key\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.362331 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25qmf\" (UniqueName: \"kubernetes.io/projected/6ac3e0e9-5501-4af6-a745-3ef391fcac43-kube-api-access-25qmf\") pod \"horizon-745f47d489-qq69f\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.498556 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.616728 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fj8mf"] Oct 02 07:32:52 crc kubenswrapper[4960]: I1002 07:32:52.800892 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-747ffb4f4f-zhxrd"] Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.057909 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-sl4rw"] Oct 02 07:32:53 crc kubenswrapper[4960]: W1002 07:32:53.058468 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8478ccab_58bf_45e2_9f7a_c5a4b0bcc216.slice/crio-d5c6b8aa307cc639b36c4b7616c31292cbae6f58156cd70752f66751d7b97583 WatchSource:0}: Error finding container d5c6b8aa307cc639b36c4b7616c31292cbae6f58156cd70752f66751d7b97583: Status 404 returned error can't find the container with id d5c6b8aa307cc639b36c4b7616c31292cbae6f58156cd70752f66751d7b97583 Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.088668 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.104621 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sl4rw" event={"ID":"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216","Type":"ContainerStarted","Data":"d5c6b8aa307cc639b36c4b7616c31292cbae6f58156cd70752f66751d7b97583"} Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.112961 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" event={"ID":"04c274e5-afed-4908-9f2e-1cd8165f9e6d","Type":"ContainerStarted","Data":"7d54cf46a138210914d5f61055b8cdc94174d162264e2aceb8565847289cc3d3"} Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.126095 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj8mf" event={"ID":"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f","Type":"ContainerStarted","Data":"d6a87fa707ede52eb6ec04e603acdde85d44b5c90e49497c2f4d6c8c805db2de"} Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.126159 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj8mf" event={"ID":"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f","Type":"ContainerStarted","Data":"c410460ea05d6ab5312fe5a4199f8149867aa268707e5a1387925cccfcab10d8"} Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.130952 4960 generic.go:334] "Generic (PLEG): container finished" podID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerID="e763f445778bfd42a76b46f2c787c94f68f8dd5a33332d5029008b8f59fa4f2b" exitCode=0 Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.131026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" event={"ID":"d58a2b49-97bb-4175-8b96-5f49fa95848b","Type":"ContainerDied","Data":"e763f445778bfd42a76b46f2c787c94f68f8dd5a33332d5029008b8f59fa4f2b"} Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.156829 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fj8mf" podStartSLOduration=2.156798024 podStartE2EDuration="2.156798024s" podCreationTimestamp="2025-10-02 07:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:32:53.151916732 +0000 UTC m=+994.183863019" watchObservedRunningTime="2025-10-02 07:32:53.156798024 +0000 UTC m=+994.188744311" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.204156 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-844dfdd985-j275s"] Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.330501 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-745f47d489-qq69f"] Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.345783 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-796bc46545-gltd7"] Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.361701 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.467936 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj72g\" (UniqueName: \"kubernetes.io/projected/d58a2b49-97bb-4175-8b96-5f49fa95848b-kube-api-access-tj72g\") pod \"d58a2b49-97bb-4175-8b96-5f49fa95848b\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.467987 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-sb\") pod \"d58a2b49-97bb-4175-8b96-5f49fa95848b\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.468086 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-dns-svc\") pod \"d58a2b49-97bb-4175-8b96-5f49fa95848b\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.468607 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-nb\") pod \"d58a2b49-97bb-4175-8b96-5f49fa95848b\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.468684 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-config\") pod \"d58a2b49-97bb-4175-8b96-5f49fa95848b\" (UID: \"d58a2b49-97bb-4175-8b96-5f49fa95848b\") " Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.473392 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d58a2b49-97bb-4175-8b96-5f49fa95848b-kube-api-access-tj72g" (OuterVolumeSpecName: "kube-api-access-tj72g") pod "d58a2b49-97bb-4175-8b96-5f49fa95848b" (UID: "d58a2b49-97bb-4175-8b96-5f49fa95848b"). InnerVolumeSpecName "kube-api-access-tj72g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.513929 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-config" (OuterVolumeSpecName: "config") pod "d58a2b49-97bb-4175-8b96-5f49fa95848b" (UID: "d58a2b49-97bb-4175-8b96-5f49fa95848b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.517674 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d58a2b49-97bb-4175-8b96-5f49fa95848b" (UID: "d58a2b49-97bb-4175-8b96-5f49fa95848b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.532708 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d58a2b49-97bb-4175-8b96-5f49fa95848b" (UID: "d58a2b49-97bb-4175-8b96-5f49fa95848b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.544855 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d58a2b49-97bb-4175-8b96-5f49fa95848b" (UID: "d58a2b49-97bb-4175-8b96-5f49fa95848b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.573720 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.573765 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj72g\" (UniqueName: \"kubernetes.io/projected/d58a2b49-97bb-4175-8b96-5f49fa95848b-kube-api-access-tj72g\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.573778 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.573790 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.573799 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a2b49-97bb-4175-8b96-5f49fa95848b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.914644 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-745f47d489-qq69f"] Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.974646 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-59867977-hxr9v"] Oct 02 07:32:53 crc kubenswrapper[4960]: E1002 07:32:53.982619 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerName="dnsmasq-dns" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.982651 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerName="dnsmasq-dns" Oct 02 07:32:53 crc kubenswrapper[4960]: E1002 07:32:53.982676 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerName="init" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.982683 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerName="init" Oct 02 07:32:53 crc kubenswrapper[4960]: I1002 07:32:53.984787 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d58a2b49-97bb-4175-8b96-5f49fa95848b" containerName="dnsmasq-dns" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.018579 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.033864 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.097023 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f5kc\" (UniqueName: \"kubernetes.io/projected/a1f5833e-dad0-46e3-b460-b6be789f3c62-kube-api-access-4f5kc\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.097132 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1f5833e-dad0-46e3-b460-b6be789f3c62-horizon-secret-key\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.097166 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f5833e-dad0-46e3-b460-b6be789f3c62-logs\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.097191 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-scripts\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.097225 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-config-data\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.103115 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59867977-hxr9v"] Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.171250 4960 generic.go:334] "Generic (PLEG): container finished" podID="04c274e5-afed-4908-9f2e-1cd8165f9e6d" containerID="69e92bed77464cb253e792cb0f71f47b0df05485abafe13a94cdcad2678d68ef" exitCode=0 Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.171510 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" event={"ID":"04c274e5-afed-4908-9f2e-1cd8165f9e6d","Type":"ContainerDied","Data":"69e92bed77464cb253e792cb0f71f47b0df05485abafe13a94cdcad2678d68ef"} Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.190639 4960 generic.go:334] "Generic (PLEG): container finished" podID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerID="e1f184fdd7a6b40030c1d8b5b059a1b54690326b9efee674466aa360b36aa949" exitCode=0 Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.190812 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-796bc46545-gltd7" event={"ID":"0a970f1b-0b01-4b2c-821b-a2c9660e1845","Type":"ContainerDied","Data":"e1f184fdd7a6b40030c1d8b5b059a1b54690326b9efee674466aa360b36aa949"} Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.190853 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-796bc46545-gltd7" event={"ID":"0a970f1b-0b01-4b2c-821b-a2c9660e1845","Type":"ContainerStarted","Data":"1cd9c40821a47d8bf6d13680ca5eb280ed6cd1ec6429e2eb48c3e58295db860d"} Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.200637 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34621790-60f4-4239-a57d-941591637b6e","Type":"ContainerStarted","Data":"b7bdf6901b4749974ad0f83e4bb54c4b671eea632a1c0780e8d6598ca273b14a"} Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.201079 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1f5833e-dad0-46e3-b460-b6be789f3c62-horizon-secret-key\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.201128 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f5833e-dad0-46e3-b460-b6be789f3c62-logs\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.201173 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-scripts\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.201215 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-config-data\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.201233 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f5kc\" (UniqueName: \"kubernetes.io/projected/a1f5833e-dad0-46e3-b460-b6be789f3c62-kube-api-access-4f5kc\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.202903 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-scripts\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.204193 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-config-data\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.205465 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-745f47d489-qq69f" event={"ID":"6ac3e0e9-5501-4af6-a745-3ef391fcac43","Type":"ContainerStarted","Data":"01aa1bb0340a8a40ee0e855bc4f6f16b6a34ab90136854430342017dac64917d"} Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.206367 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f5833e-dad0-46e3-b460-b6be789f3c62-logs\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.211658 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1f5833e-dad0-46e3-b460-b6be789f3c62-horizon-secret-key\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.218860 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" event={"ID":"d58a2b49-97bb-4175-8b96-5f49fa95848b","Type":"ContainerDied","Data":"78e1aecd8165e0ad141bedac2154abf83aa125e769acf0e0dd93d3652d3bdf9a"} Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.218966 4960 scope.go:117] "RemoveContainer" containerID="e763f445778bfd42a76b46f2c787c94f68f8dd5a33332d5029008b8f59fa4f2b" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.219229 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9dd498fbf-fgq6l" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.223279 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f5kc\" (UniqueName: \"kubernetes.io/projected/a1f5833e-dad0-46e3-b460-b6be789f3c62-kube-api-access-4f5kc\") pod \"horizon-59867977-hxr9v\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.247454 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-844dfdd985-j275s" event={"ID":"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75","Type":"ContainerStarted","Data":"9243b74783bdacd0ea21732ccc710997840cf6d8e98fe73cddcd74e3c3f00322"} Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.341244 4960 scope.go:117] "RemoveContainer" containerID="941e3986ce90567e2192200768c0e7bcd3ab1a6a6c75ad93f1e48dceb1d275ed" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.355735 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9dd498fbf-fgq6l"] Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.357459 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9dd498fbf-fgq6l"] Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.367479 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59867977-hxr9v" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.828443 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.926128 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2958\" (UniqueName: \"kubernetes.io/projected/04c274e5-afed-4908-9f2e-1cd8165f9e6d-kube-api-access-k2958\") pod \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.926796 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-nb\") pod \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.926892 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-sb\") pod \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.926976 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-config\") pod \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.927054 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-dns-svc\") pod \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\" (UID: \"04c274e5-afed-4908-9f2e-1cd8165f9e6d\") " Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.938672 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04c274e5-afed-4908-9f2e-1cd8165f9e6d-kube-api-access-k2958" (OuterVolumeSpecName: "kube-api-access-k2958") pod "04c274e5-afed-4908-9f2e-1cd8165f9e6d" (UID: "04c274e5-afed-4908-9f2e-1cd8165f9e6d"). InnerVolumeSpecName "kube-api-access-k2958". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.971309 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "04c274e5-afed-4908-9f2e-1cd8165f9e6d" (UID: "04c274e5-afed-4908-9f2e-1cd8165f9e6d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:54 crc kubenswrapper[4960]: I1002 07:32:54.981502 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "04c274e5-afed-4908-9f2e-1cd8165f9e6d" (UID: "04c274e5-afed-4908-9f2e-1cd8165f9e6d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.006743 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-config" (OuterVolumeSpecName: "config") pod "04c274e5-afed-4908-9f2e-1cd8165f9e6d" (UID: "04c274e5-afed-4908-9f2e-1cd8165f9e6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.015043 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "04c274e5-afed-4908-9f2e-1cd8165f9e6d" (UID: "04c274e5-afed-4908-9f2e-1cd8165f9e6d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.030591 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2958\" (UniqueName: \"kubernetes.io/projected/04c274e5-afed-4908-9f2e-1cd8165f9e6d-kube-api-access-k2958\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.030638 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.030651 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.030664 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.030679 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04c274e5-afed-4908-9f2e-1cd8165f9e6d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.253726 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59867977-hxr9v"] Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.262585 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-796bc46545-gltd7" event={"ID":"0a970f1b-0b01-4b2c-821b-a2c9660e1845","Type":"ContainerStarted","Data":"ae02f767c7cd874afd138f88d701ca1a8e2cf589d376756c330d330779956b0a"} Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.263046 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:32:55 crc kubenswrapper[4960]: W1002 07:32:55.271692 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1f5833e_dad0_46e3_b460_b6be789f3c62.slice/crio-360fe37038f042a1e95da8bd4c8166ffd9535b5364a2708ea07b72210d4100fd WatchSource:0}: Error finding container 360fe37038f042a1e95da8bd4c8166ffd9535b5364a2708ea07b72210d4100fd: Status 404 returned error can't find the container with id 360fe37038f042a1e95da8bd4c8166ffd9535b5364a2708ea07b72210d4100fd Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.277313 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" event={"ID":"04c274e5-afed-4908-9f2e-1cd8165f9e6d","Type":"ContainerDied","Data":"7d54cf46a138210914d5f61055b8cdc94174d162264e2aceb8565847289cc3d3"} Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.277404 4960 scope.go:117] "RemoveContainer" containerID="69e92bed77464cb253e792cb0f71f47b0df05485abafe13a94cdcad2678d68ef" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.277538 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-747ffb4f4f-zhxrd" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.304929 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-796bc46545-gltd7" podStartSLOduration=4.304907311 podStartE2EDuration="4.304907311s" podCreationTimestamp="2025-10-02 07:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:32:55.296302452 +0000 UTC m=+996.328248739" watchObservedRunningTime="2025-10-02 07:32:55.304907311 +0000 UTC m=+996.336853598" Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.361020 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-747ffb4f4f-zhxrd"] Oct 02 07:32:55 crc kubenswrapper[4960]: I1002 07:32:55.367298 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-747ffb4f4f-zhxrd"] Oct 02 07:32:56 crc kubenswrapper[4960]: I1002 07:32:56.301746 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59867977-hxr9v" event={"ID":"a1f5833e-dad0-46e3-b460-b6be789f3c62","Type":"ContainerStarted","Data":"360fe37038f042a1e95da8bd4c8166ffd9535b5364a2708ea07b72210d4100fd"} Oct 02 07:32:56 crc kubenswrapper[4960]: I1002 07:32:56.345398 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04c274e5-afed-4908-9f2e-1cd8165f9e6d" path="/var/lib/kubelet/pods/04c274e5-afed-4908-9f2e-1cd8165f9e6d/volumes" Oct 02 07:32:56 crc kubenswrapper[4960]: I1002 07:32:56.346492 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d58a2b49-97bb-4175-8b96-5f49fa95848b" path="/var/lib/kubelet/pods/d58a2b49-97bb-4175-8b96-5f49fa95848b/volumes" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.767802 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a856-account-create-h8wpl"] Oct 02 07:32:58 crc kubenswrapper[4960]: E1002 07:32:58.768617 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04c274e5-afed-4908-9f2e-1cd8165f9e6d" containerName="init" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.768631 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="04c274e5-afed-4908-9f2e-1cd8165f9e6d" containerName="init" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.768798 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="04c274e5-afed-4908-9f2e-1cd8165f9e6d" containerName="init" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.769495 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a856-account-create-h8wpl" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.772281 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.781371 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a856-account-create-h8wpl"] Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.833543 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6dfq\" (UniqueName: \"kubernetes.io/projected/5eb7e0fd-eb02-4f1a-af2f-d61e09104130-kube-api-access-k6dfq\") pod \"barbican-a856-account-create-h8wpl\" (UID: \"5eb7e0fd-eb02-4f1a-af2f-d61e09104130\") " pod="openstack/barbican-a856-account-create-h8wpl" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.873749 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-de0b-account-create-6q6nn"] Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.875309 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de0b-account-create-6q6nn" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.877708 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.887337 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-de0b-account-create-6q6nn"] Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.936036 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5kws\" (UniqueName: \"kubernetes.io/projected/781f3c80-6287-411d-9c26-ab52d38c8291-kube-api-access-k5kws\") pod \"cinder-de0b-account-create-6q6nn\" (UID: \"781f3c80-6287-411d-9c26-ab52d38c8291\") " pod="openstack/cinder-de0b-account-create-6q6nn" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.936197 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6dfq\" (UniqueName: \"kubernetes.io/projected/5eb7e0fd-eb02-4f1a-af2f-d61e09104130-kube-api-access-k6dfq\") pod \"barbican-a856-account-create-h8wpl\" (UID: \"5eb7e0fd-eb02-4f1a-af2f-d61e09104130\") " pod="openstack/barbican-a856-account-create-h8wpl" Oct 02 07:32:58 crc kubenswrapper[4960]: I1002 07:32:58.968764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6dfq\" (UniqueName: \"kubernetes.io/projected/5eb7e0fd-eb02-4f1a-af2f-d61e09104130-kube-api-access-k6dfq\") pod \"barbican-a856-account-create-h8wpl\" (UID: \"5eb7e0fd-eb02-4f1a-af2f-d61e09104130\") " pod="openstack/barbican-a856-account-create-h8wpl" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.038581 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5kws\" (UniqueName: \"kubernetes.io/projected/781f3c80-6287-411d-9c26-ab52d38c8291-kube-api-access-k5kws\") pod \"cinder-de0b-account-create-6q6nn\" (UID: \"781f3c80-6287-411d-9c26-ab52d38c8291\") " pod="openstack/cinder-de0b-account-create-6q6nn" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.067262 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5kws\" (UniqueName: \"kubernetes.io/projected/781f3c80-6287-411d-9c26-ab52d38c8291-kube-api-access-k5kws\") pod \"cinder-de0b-account-create-6q6nn\" (UID: \"781f3c80-6287-411d-9c26-ab52d38c8291\") " pod="openstack/cinder-de0b-account-create-6q6nn" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.078284 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b4f0-account-create-sq74s"] Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.079906 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b4f0-account-create-sq74s" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.082686 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.100357 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a856-account-create-h8wpl" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.110121 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b4f0-account-create-sq74s"] Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.140936 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlpwm\" (UniqueName: \"kubernetes.io/projected/f7fa3598-b2f0-4a29-9150-0b881db98bcb-kube-api-access-wlpwm\") pod \"neutron-b4f0-account-create-sq74s\" (UID: \"f7fa3598-b2f0-4a29-9150-0b881db98bcb\") " pod="openstack/neutron-b4f0-account-create-sq74s" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.194458 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de0b-account-create-6q6nn" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.243753 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlpwm\" (UniqueName: \"kubernetes.io/projected/f7fa3598-b2f0-4a29-9150-0b881db98bcb-kube-api-access-wlpwm\") pod \"neutron-b4f0-account-create-sq74s\" (UID: \"f7fa3598-b2f0-4a29-9150-0b881db98bcb\") " pod="openstack/neutron-b4f0-account-create-sq74s" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.262105 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlpwm\" (UniqueName: \"kubernetes.io/projected/f7fa3598-b2f0-4a29-9150-0b881db98bcb-kube-api-access-wlpwm\") pod \"neutron-b4f0-account-create-sq74s\" (UID: \"f7fa3598-b2f0-4a29-9150-0b881db98bcb\") " pod="openstack/neutron-b4f0-account-create-sq74s" Oct 02 07:32:59 crc kubenswrapper[4960]: I1002 07:32:59.445454 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b4f0-account-create-sq74s" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.374413 4960 generic.go:334] "Generic (PLEG): container finished" podID="8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" containerID="d6a87fa707ede52eb6ec04e603acdde85d44b5c90e49497c2f4d6c8c805db2de" exitCode=0 Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.374470 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj8mf" event={"ID":"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f","Type":"ContainerDied","Data":"d6a87fa707ede52eb6ec04e603acdde85d44b5c90e49497c2f4d6c8c805db2de"} Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.765966 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-844dfdd985-j275s"] Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.797331 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85f9f4c9c8-9kk6b"] Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.798813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.803319 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.825886 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85f9f4c9c8-9kk6b"] Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.872928 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59867977-hxr9v"] Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.882739 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-scripts\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.882821 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-tls-certs\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.882856 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-secret-key\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.882900 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12495902-a857-4ea0-b5bc-178803f33c12-logs\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.882928 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-combined-ca-bundle\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.882994 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-config-data\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.883017 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvsmj\" (UniqueName: \"kubernetes.io/projected/12495902-a857-4ea0-b5bc-178803f33c12-kube-api-access-qvsmj\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.894989 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d66c446d8-rxzk8"] Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.896566 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.930390 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d66c446d8-rxzk8"] Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.984377 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35082c42-a42f-4eb7-a12c-91c74a6715f4-logs\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.984807 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12495902-a857-4ea0-b5bc-178803f33c12-logs\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.984853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-combined-ca-bundle\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.984879 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j68jg\" (UniqueName: \"kubernetes.io/projected/35082c42-a42f-4eb7-a12c-91c74a6715f4-kube-api-access-j68jg\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.984903 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-combined-ca-bundle\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.984960 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-config-data\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985104 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvsmj\" (UniqueName: \"kubernetes.io/projected/12495902-a857-4ea0-b5bc-178803f33c12-kube-api-access-qvsmj\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985269 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12495902-a857-4ea0-b5bc-178803f33c12-logs\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985291 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-horizon-tls-certs\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-scripts\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985424 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-horizon-secret-key\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985472 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35082c42-a42f-4eb7-a12c-91c74a6715f4-scripts\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985590 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-tls-certs\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985623 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35082c42-a42f-4eb7-a12c-91c74a6715f4-config-data\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.985709 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-secret-key\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.986763 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-config-data\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.987004 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-scripts\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.991944 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-combined-ca-bundle\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.993604 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-tls-certs\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:00 crc kubenswrapper[4960]: I1002 07:33:00.994938 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-secret-key\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.004075 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvsmj\" (UniqueName: \"kubernetes.io/projected/12495902-a857-4ea0-b5bc-178803f33c12-kube-api-access-qvsmj\") pod \"horizon-85f9f4c9c8-9kk6b\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.087136 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-horizon-tls-certs\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.087183 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-horizon-secret-key\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.087206 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35082c42-a42f-4eb7-a12c-91c74a6715f4-scripts\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.087242 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35082c42-a42f-4eb7-a12c-91c74a6715f4-config-data\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.087286 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35082c42-a42f-4eb7-a12c-91c74a6715f4-logs\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.087322 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j68jg\" (UniqueName: \"kubernetes.io/projected/35082c42-a42f-4eb7-a12c-91c74a6715f4-kube-api-access-j68jg\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.087363 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-combined-ca-bundle\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.088160 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35082c42-a42f-4eb7-a12c-91c74a6715f4-scripts\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.089631 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35082c42-a42f-4eb7-a12c-91c74a6715f4-logs\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.090921 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35082c42-a42f-4eb7-a12c-91c74a6715f4-config-data\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.094666 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-horizon-tls-certs\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.094934 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-horizon-secret-key\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.100767 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35082c42-a42f-4eb7-a12c-91c74a6715f4-combined-ca-bundle\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.112493 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j68jg\" (UniqueName: \"kubernetes.io/projected/35082c42-a42f-4eb7-a12c-91c74a6715f4-kube-api-access-j68jg\") pod \"horizon-d66c446d8-rxzk8\" (UID: \"35082c42-a42f-4eb7-a12c-91c74a6715f4\") " pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.123836 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:01 crc kubenswrapper[4960]: I1002 07:33:01.219864 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:02 crc kubenswrapper[4960]: I1002 07:33:02.351349 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:33:02 crc kubenswrapper[4960]: I1002 07:33:02.412719 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dd86f89c-bppxm"] Oct 02 07:33:02 crc kubenswrapper[4960]: I1002 07:33:02.415466 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="dnsmasq-dns" containerID="cri-o://89b90dec3c5e437f50f3c08f10cc2f3aef20852c15382cb33db6ad69795db39a" gracePeriod=10 Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.356220 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.425832 4960 generic.go:334] "Generic (PLEG): container finished" podID="a13a796e-8fc2-463c-9587-8fc19145de29" containerID="89b90dec3c5e437f50f3c08f10cc2f3aef20852c15382cb33db6ad69795db39a" exitCode=0 Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.425893 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" event={"ID":"a13a796e-8fc2-463c-9587-8fc19145de29","Type":"ContainerDied","Data":"89b90dec3c5e437f50f3c08f10cc2f3aef20852c15382cb33db6ad69795db39a"} Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.875945 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.956491 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-config-data\") pod \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.956598 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-combined-ca-bundle\") pod \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.956669 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-scripts\") pod \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.956825 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kxxj\" (UniqueName: \"kubernetes.io/projected/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-kube-api-access-7kxxj\") pod \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.956876 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-fernet-keys\") pod \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.957028 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-credential-keys\") pod \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\" (UID: \"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f\") " Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.968162 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-scripts" (OuterVolumeSpecName: "scripts") pod "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" (UID: "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.969262 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-kube-api-access-7kxxj" (OuterVolumeSpecName: "kube-api-access-7kxxj") pod "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" (UID: "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f"). InnerVolumeSpecName "kube-api-access-7kxxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.983085 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" (UID: "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.991891 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" (UID: "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.995148 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" (UID: "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4960]: I1002 07:33:03.998863 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-config-data" (OuterVolumeSpecName: "config-data") pod "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" (UID: "8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.059739 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kxxj\" (UniqueName: \"kubernetes.io/projected/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-kube-api-access-7kxxj\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.059793 4960 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.059805 4960 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.059815 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.059824 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.059832 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.438128 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj8mf" event={"ID":"8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f","Type":"ContainerDied","Data":"c410460ea05d6ab5312fe5a4199f8149867aa268707e5a1387925cccfcab10d8"} Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.438185 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c410460ea05d6ab5312fe5a4199f8149867aa268707e5a1387925cccfcab10d8" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.438255 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj8mf" Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.950131 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fj8mf"] Oct 02 07:33:04 crc kubenswrapper[4960]: I1002 07:33:04.959634 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fj8mf"] Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.072339 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xjxcq"] Oct 02 07:33:05 crc kubenswrapper[4960]: E1002 07:33:05.073057 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" containerName="keystone-bootstrap" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.073118 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" containerName="keystone-bootstrap" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.073621 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" containerName="keystone-bootstrap" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.074803 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.079190 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.079645 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mplxm" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.079728 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.079735 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.100481 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xjxcq"] Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.194031 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-combined-ca-bundle\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.194093 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlf9s\" (UniqueName: \"kubernetes.io/projected/14972b11-92c6-4bea-bf0e-f926c18c198f-kube-api-access-xlf9s\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.194157 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-config-data\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.194198 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-credential-keys\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.194223 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-scripts\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.194254 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-fernet-keys\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.295918 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-config-data\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.296029 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-credential-keys\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.296057 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-scripts\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.296091 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-fernet-keys\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.296156 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-combined-ca-bundle\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.296181 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlf9s\" (UniqueName: \"kubernetes.io/projected/14972b11-92c6-4bea-bf0e-f926c18c198f-kube-api-access-xlf9s\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.304095 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-scripts\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.304436 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-fernet-keys\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.304528 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-config-data\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.306660 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-combined-ca-bundle\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.314914 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-credential-keys\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.315398 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlf9s\" (UniqueName: \"kubernetes.io/projected/14972b11-92c6-4bea-bf0e-f926c18c198f-kube-api-access-xlf9s\") pod \"keystone-bootstrap-xjxcq\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:05 crc kubenswrapper[4960]: I1002 07:33:05.402574 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:06 crc kubenswrapper[4960]: I1002 07:33:06.345377 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f" path="/var/lib/kubelet/pods/8405d0c8-63c7-4c98-ab5c-9265a7eb7d5f/volumes" Oct 02 07:33:08 crc kubenswrapper[4960]: I1002 07:33:08.357147 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 02 07:33:09 crc kubenswrapper[4960]: E1002 07:33:09.826767 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:cd4347fc6be340bb0a5d1fa5046f0372d5a2193d4913ed1e8038b4070fae5169" Oct 02 07:33:09 crc kubenswrapper[4960]: E1002 07:33:09.827149 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:cd4347fc6be340bb0a5d1fa5046f0372d5a2193d4913ed1e8038b4070fae5169,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5cdh4h696h57bh55h558h8ch57h65h576h65bhb8h54dh676hf9h588h8ch65bh5f4h569h5fch657h676h5f4h64dhf4h6fh674hbbh9bhb6h78q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-25qmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-745f47d489-qq69f_openstack(6ac3e0e9-5501-4af6-a745-3ef391fcac43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:33:09 crc kubenswrapper[4960]: E1002 07:33:09.828702 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:cd4347fc6be340bb0a5d1fa5046f0372d5a2193d4913ed1e8038b4070fae5169" Oct 02 07:33:09 crc kubenswrapper[4960]: E1002 07:33:09.828878 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:cd4347fc6be340bb0a5d1fa5046f0372d5a2193d4913ed1e8038b4070fae5169,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5f7h656h57bh5d6h585h57chd7h5ch649h59bhbch5f6h588h579hb4h685hd7h8fh655h548h647h85h648hc5h5ddh584h69h698hb9h84h5ffh56bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pbczx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-844dfdd985-j275s_openstack(7f4fe04f-34d3-4db7-a4cd-e34a28d63d75): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:33:09 crc kubenswrapper[4960]: E1002 07:33:09.836465 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:cd4347fc6be340bb0a5d1fa5046f0372d5a2193d4913ed1e8038b4070fae5169\\\"\"]" pod="openstack/horizon-844dfdd985-j275s" podUID="7f4fe04f-34d3-4db7-a4cd-e34a28d63d75" Oct 02 07:33:09 crc kubenswrapper[4960]: E1002 07:33:09.836646 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:cd4347fc6be340bb0a5d1fa5046f0372d5a2193d4913ed1e8038b4070fae5169\\\"\"]" pod="openstack/horizon-745f47d489-qq69f" podUID="6ac3e0e9-5501-4af6-a745-3ef391fcac43" Oct 02 07:33:10 crc kubenswrapper[4960]: E1002 07:33:10.285523 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:17a5a8c217d3344d7fbe8dd1b01a4a99194aad59c2683ccd39d077f528c0880e" Oct 02 07:33:10 crc kubenswrapper[4960]: E1002 07:33:10.286078 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:17a5a8c217d3344d7fbe8dd1b01a4a99194aad59c2683ccd39d077f528c0880e,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n677h5c5h9fh89hc6h5f7h87hc5h67dh8dh89h6ch5b4hcch78hchf9h584h54dh6bh89h67ch687h578h659h56dh5fh9hcbh5bfh685h5fbq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lxqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(34621790-60f4-4239-a57d-941591637b6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.601913 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.735017 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb\") pod \"a13a796e-8fc2-463c-9587-8fc19145de29\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.735411 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-config\") pod \"a13a796e-8fc2-463c-9587-8fc19145de29\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.735753 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5mzx\" (UniqueName: \"kubernetes.io/projected/a13a796e-8fc2-463c-9587-8fc19145de29-kube-api-access-s5mzx\") pod \"a13a796e-8fc2-463c-9587-8fc19145de29\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.735875 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-sb\") pod \"a13a796e-8fc2-463c-9587-8fc19145de29\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.735925 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-dns-svc\") pod \"a13a796e-8fc2-463c-9587-8fc19145de29\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.747839 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a13a796e-8fc2-463c-9587-8fc19145de29-kube-api-access-s5mzx" (OuterVolumeSpecName: "kube-api-access-s5mzx") pod "a13a796e-8fc2-463c-9587-8fc19145de29" (UID: "a13a796e-8fc2-463c-9587-8fc19145de29"). InnerVolumeSpecName "kube-api-access-s5mzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.806703 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-config" (OuterVolumeSpecName: "config") pod "a13a796e-8fc2-463c-9587-8fc19145de29" (UID: "a13a796e-8fc2-463c-9587-8fc19145de29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.806842 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a13a796e-8fc2-463c-9587-8fc19145de29" (UID: "a13a796e-8fc2-463c-9587-8fc19145de29"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4960]: E1002 07:33:10.830518 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb podName:a13a796e-8fc2-463c-9587-8fc19145de29 nodeName:}" failed. No retries permitted until 2025-10-02 07:33:11.330471606 +0000 UTC m=+1012.362417883 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb") pod "a13a796e-8fc2-463c-9587-8fc19145de29" (UID: "a13a796e-8fc2-463c-9587-8fc19145de29") : error deleting /var/lib/kubelet/pods/a13a796e-8fc2-463c-9587-8fc19145de29/volume-subpaths: remove /var/lib/kubelet/pods/a13a796e-8fc2-463c-9587-8fc19145de29/volume-subpaths: no such file or directory Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.830840 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a13a796e-8fc2-463c-9587-8fc19145de29" (UID: "a13a796e-8fc2-463c-9587-8fc19145de29"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.839167 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.839210 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5mzx\" (UniqueName: \"kubernetes.io/projected/a13a796e-8fc2-463c-9587-8fc19145de29-kube-api-access-s5mzx\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.839227 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:10 crc kubenswrapper[4960]: I1002 07:33:10.839237 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.150226 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.164294 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.230717 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-de0b-account-create-6q6nn"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.248846 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbczx\" (UniqueName: \"kubernetes.io/projected/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-kube-api-access-pbczx\") pod \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.248919 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-config-data\") pod \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.249076 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-horizon-secret-key\") pod \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.249114 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-scripts\") pod \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.249151 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-logs\") pod \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\" (UID: \"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.249848 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-logs" (OuterVolumeSpecName: "logs") pod "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75" (UID: "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.249960 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b4f0-account-create-sq74s"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.250080 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-config-data" (OuterVolumeSpecName: "config-data") pod "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75" (UID: "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.250185 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-scripts" (OuterVolumeSpecName: "scripts") pod "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75" (UID: "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.254793 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-kube-api-access-pbczx" (OuterVolumeSpecName: "kube-api-access-pbczx") pod "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75" (UID: "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75"). InnerVolumeSpecName "kube-api-access-pbczx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.254873 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75" (UID: "7f4fe04f-34d3-4db7-a4cd-e34a28d63d75"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: W1002 07:33:11.256424 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7fa3598_b2f0_4a29_9150_0b881db98bcb.slice/crio-856c1da669a21f3d26e9c78326b80309541b60c7512b1c466bc919aa1f7a0bea WatchSource:0}: Error finding container 856c1da669a21f3d26e9c78326b80309541b60c7512b1c466bc919aa1f7a0bea: Status 404 returned error can't find the container with id 856c1da669a21f3d26e9c78326b80309541b60c7512b1c466bc919aa1f7a0bea Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.352722 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ac3e0e9-5501-4af6-a745-3ef391fcac43-horizon-secret-key\") pod \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.352788 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-config-data\") pod \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.352831 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb\") pod \"a13a796e-8fc2-463c-9587-8fc19145de29\" (UID: \"a13a796e-8fc2-463c-9587-8fc19145de29\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.352867 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-scripts\") pod \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.353079 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac3e0e9-5501-4af6-a745-3ef391fcac43-logs\") pod \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.353138 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25qmf\" (UniqueName: \"kubernetes.io/projected/6ac3e0e9-5501-4af6-a745-3ef391fcac43-kube-api-access-25qmf\") pod \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\" (UID: \"6ac3e0e9-5501-4af6-a745-3ef391fcac43\") " Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.353386 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-scripts" (OuterVolumeSpecName: "scripts") pod "6ac3e0e9-5501-4af6-a745-3ef391fcac43" (UID: "6ac3e0e9-5501-4af6-a745-3ef391fcac43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.353499 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-config-data" (OuterVolumeSpecName: "config-data") pod "6ac3e0e9-5501-4af6-a745-3ef391fcac43" (UID: "6ac3e0e9-5501-4af6-a745-3ef391fcac43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.353708 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac3e0e9-5501-4af6-a745-3ef391fcac43-logs" (OuterVolumeSpecName: "logs") pod "6ac3e0e9-5501-4af6-a745-3ef391fcac43" (UID: "6ac3e0e9-5501-4af6-a745-3ef391fcac43"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.353393 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a13a796e-8fc2-463c-9587-8fc19145de29" (UID: "a13a796e-8fc2-463c-9587-8fc19145de29"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354179 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354205 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a13a796e-8fc2-463c-9587-8fc19145de29-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354217 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ac3e0e9-5501-4af6-a745-3ef391fcac43-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354228 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbczx\" (UniqueName: \"kubernetes.io/projected/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-kube-api-access-pbczx\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354238 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354247 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac3e0e9-5501-4af6-a745-3ef391fcac43-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354256 4960 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354267 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.354275 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.358145 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac3e0e9-5501-4af6-a745-3ef391fcac43-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6ac3e0e9-5501-4af6-a745-3ef391fcac43" (UID: "6ac3e0e9-5501-4af6-a745-3ef391fcac43"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.358338 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac3e0e9-5501-4af6-a745-3ef391fcac43-kube-api-access-25qmf" (OuterVolumeSpecName: "kube-api-access-25qmf") pod "6ac3e0e9-5501-4af6-a745-3ef391fcac43" (UID: "6ac3e0e9-5501-4af6-a745-3ef391fcac43"). InnerVolumeSpecName "kube-api-access-25qmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.461178 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25qmf\" (UniqueName: \"kubernetes.io/projected/6ac3e0e9-5501-4af6-a745-3ef391fcac43-kube-api-access-25qmf\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.461216 4960 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ac3e0e9-5501-4af6-a745-3ef391fcac43-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.465403 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a856-account-create-h8wpl"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.471438 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xjxcq"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.495701 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d66c446d8-rxzk8"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.502176 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85f9f4c9c8-9kk6b"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.538055 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-745f47d489-qq69f" event={"ID":"6ac3e0e9-5501-4af6-a745-3ef391fcac43","Type":"ContainerDied","Data":"01aa1bb0340a8a40ee0e855bc4f6f16b6a34ab90136854430342017dac64917d"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.538145 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-745f47d489-qq69f" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.542993 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b4f0-account-create-sq74s" event={"ID":"f7fa3598-b2f0-4a29-9150-0b881db98bcb","Type":"ContainerStarted","Data":"8bb13f6369399e014df1695dbe725e4e2eb136bc25f0b6d74a1833373be25b2a"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.543032 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b4f0-account-create-sq74s" event={"ID":"f7fa3598-b2f0-4a29-9150-0b881db98bcb","Type":"ContainerStarted","Data":"856c1da669a21f3d26e9c78326b80309541b60c7512b1c466bc919aa1f7a0bea"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.545296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sl4rw" event={"ID":"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216","Type":"ContainerStarted","Data":"a495c74eb425cc66ed07cdc3565ac92fb0743250ceb3c81f3a4f528f0697997a"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.548705 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-de0b-account-create-6q6nn" event={"ID":"781f3c80-6287-411d-9c26-ab52d38c8291","Type":"ContainerStarted","Data":"ebae7086f06dcac78a6692cf60e174260e1fee556e1398b0e38064bd0aab6a51"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.548727 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-de0b-account-create-6q6nn" event={"ID":"781f3c80-6287-411d-9c26-ab52d38c8291","Type":"ContainerStarted","Data":"2dcc735891037adb25eb8458251ad3475df64e645214af36c94b6f023d73b623"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.559570 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a856-account-create-h8wpl" event={"ID":"5eb7e0fd-eb02-4f1a-af2f-d61e09104130","Type":"ContainerStarted","Data":"3cbdb4183fb1808f15e9a04cb2f69aff0c4c629ad82d161628bc4fecf6b6f6f1"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.567051 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" event={"ID":"a13a796e-8fc2-463c-9587-8fc19145de29","Type":"ContainerDied","Data":"e22d9fc4fde2defb3bc57de0f24c6f9cb47e705814fa88a65923bf210ebf8532"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.567220 4960 scope.go:117] "RemoveContainer" containerID="89b90dec3c5e437f50f3c08f10cc2f3aef20852c15382cb33db6ad69795db39a" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.567270 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dd86f89c-bppxm" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.572687 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-844dfdd985-j275s" event={"ID":"7f4fe04f-34d3-4db7-a4cd-e34a28d63d75","Type":"ContainerDied","Data":"9243b74783bdacd0ea21732ccc710997840cf6d8e98fe73cddcd74e3c3f00322"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.572786 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-844dfdd985-j275s" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.582260 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59867977-hxr9v" event={"ID":"a1f5833e-dad0-46e3-b460-b6be789f3c62","Type":"ContainerStarted","Data":"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.582308 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59867977-hxr9v" event={"ID":"a1f5833e-dad0-46e3-b460-b6be789f3c62","Type":"ContainerStarted","Data":"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.582464 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-59867977-hxr9v" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon-log" containerID="cri-o://ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc" gracePeriod=30 Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.582843 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-59867977-hxr9v" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon" containerID="cri-o://fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2" gracePeriod=30 Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.588009 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xjxcq" event={"ID":"14972b11-92c6-4bea-bf0e-f926c18c198f","Type":"ContainerStarted","Data":"c1304378d67c72a51dc58fcdef28b6eb7fcf61469dcb3d43deac6d88d9363630"} Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.602620 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-sl4rw" podStartSLOduration=3.281025882 podStartE2EDuration="20.602602293s" podCreationTimestamp="2025-10-02 07:32:51 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.063137608 +0000 UTC m=+994.095083895" lastFinishedPulling="2025-10-02 07:33:10.384714019 +0000 UTC m=+1011.416660306" observedRunningTime="2025-10-02 07:33:11.600593314 +0000 UTC m=+1012.632539601" watchObservedRunningTime="2025-10-02 07:33:11.602602293 +0000 UTC m=+1012.634548580" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.629038 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-59867977-hxr9v" podStartSLOduration=3.261079794 podStartE2EDuration="18.629012596s" podCreationTimestamp="2025-10-02 07:32:53 +0000 UTC" firstStartedPulling="2025-10-02 07:32:55.304454677 +0000 UTC m=+996.336400964" lastFinishedPulling="2025-10-02 07:33:10.672387479 +0000 UTC m=+1011.704333766" observedRunningTime="2025-10-02 07:33:11.620221631 +0000 UTC m=+1012.652167918" watchObservedRunningTime="2025-10-02 07:33:11.629012596 +0000 UTC m=+1012.660958883" Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.689140 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dd86f89c-bppxm"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.699385 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64dd86f89c-bppxm"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.722831 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-745f47d489-qq69f"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.731252 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-745f47d489-qq69f"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.746239 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-844dfdd985-j275s"] Oct 02 07:33:11 crc kubenswrapper[4960]: I1002 07:33:11.751563 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-844dfdd985-j275s"] Oct 02 07:33:12 crc kubenswrapper[4960]: W1002 07:33:12.008493 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35082c42_a42f_4eb7_a12c_91c74a6715f4.slice/crio-dd9af7612733dad9c017a64fcd040eba35f4ef60ce909caeffacc0c0e40dcc3e WatchSource:0}: Error finding container dd9af7612733dad9c017a64fcd040eba35f4ef60ce909caeffacc0c0e40dcc3e: Status 404 returned error can't find the container with id dd9af7612733dad9c017a64fcd040eba35f4ef60ce909caeffacc0c0e40dcc3e Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.049841 4960 scope.go:117] "RemoveContainer" containerID="c690443620a87e2afc239b0c63ed3e580d9860c8474aa2074a74e3008efb379a" Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.351803 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac3e0e9-5501-4af6-a745-3ef391fcac43" path="/var/lib/kubelet/pods/6ac3e0e9-5501-4af6-a745-3ef391fcac43/volumes" Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.353232 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f4fe04f-34d3-4db7-a4cd-e34a28d63d75" path="/var/lib/kubelet/pods/7f4fe04f-34d3-4db7-a4cd-e34a28d63d75/volumes" Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.353893 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" path="/var/lib/kubelet/pods/a13a796e-8fc2-463c-9587-8fc19145de29/volumes" Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.615131 4960 generic.go:334] "Generic (PLEG): container finished" podID="781f3c80-6287-411d-9c26-ab52d38c8291" containerID="ebae7086f06dcac78a6692cf60e174260e1fee556e1398b0e38064bd0aab6a51" exitCode=0 Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.615204 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-de0b-account-create-6q6nn" event={"ID":"781f3c80-6287-411d-9c26-ab52d38c8291","Type":"ContainerDied","Data":"ebae7086f06dcac78a6692cf60e174260e1fee556e1398b0e38064bd0aab6a51"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.626751 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xjxcq" event={"ID":"14972b11-92c6-4bea-bf0e-f926c18c198f","Type":"ContainerStarted","Data":"95932ea4bc707c32ada6f68dd3182b3dd177a3928dce54959f6136978a031a5f"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.635405 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34621790-60f4-4239-a57d-941591637b6e","Type":"ContainerStarted","Data":"5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.637448 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f9f4c9c8-9kk6b" event={"ID":"12495902-a857-4ea0-b5bc-178803f33c12","Type":"ContainerStarted","Data":"18bae37076d32c0f6e54e8ed4eb2151d38367a0781878ae2c2c5ffe57acd4e5f"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.637482 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f9f4c9c8-9kk6b" event={"ID":"12495902-a857-4ea0-b5bc-178803f33c12","Type":"ContainerStarted","Data":"84a535d2f7b6fb0a202941ae86960bc22df01843f5041f77fb8b72f85460edc8"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.651120 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xjxcq" podStartSLOduration=7.651079333 podStartE2EDuration="7.651079333s" podCreationTimestamp="2025-10-02 07:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:12.647437927 +0000 UTC m=+1013.679384214" watchObservedRunningTime="2025-10-02 07:33:12.651079333 +0000 UTC m=+1013.683025620" Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.656422 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d66c446d8-rxzk8" event={"ID":"35082c42-a42f-4eb7-a12c-91c74a6715f4","Type":"ContainerStarted","Data":"ad67a036237758f440da459ee6d8f8dbc70792e73ca7601484ec1ed1c3bddf97"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.656500 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d66c446d8-rxzk8" event={"ID":"35082c42-a42f-4eb7-a12c-91c74a6715f4","Type":"ContainerStarted","Data":"dd9af7612733dad9c017a64fcd040eba35f4ef60ce909caeffacc0c0e40dcc3e"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.666672 4960 generic.go:334] "Generic (PLEG): container finished" podID="5eb7e0fd-eb02-4f1a-af2f-d61e09104130" containerID="76c3a4b0776c2786b3d0129588524cc3d5c3ad291da22aa7fed44e0081b1acf3" exitCode=0 Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.666805 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a856-account-create-h8wpl" event={"ID":"5eb7e0fd-eb02-4f1a-af2f-d61e09104130","Type":"ContainerDied","Data":"76c3a4b0776c2786b3d0129588524cc3d5c3ad291da22aa7fed44e0081b1acf3"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.669752 4960 generic.go:334] "Generic (PLEG): container finished" podID="f7fa3598-b2f0-4a29-9150-0b881db98bcb" containerID="8bb13f6369399e014df1695dbe725e4e2eb136bc25f0b6d74a1833373be25b2a" exitCode=0 Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.670378 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b4f0-account-create-sq74s" event={"ID":"f7fa3598-b2f0-4a29-9150-0b881db98bcb","Type":"ContainerDied","Data":"8bb13f6369399e014df1695dbe725e4e2eb136bc25f0b6d74a1833373be25b2a"} Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.693069 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-85f9f4c9c8-9kk6b" podStartSLOduration=12.693046945 podStartE2EDuration="12.693046945s" podCreationTimestamp="2025-10-02 07:33:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:12.673029826 +0000 UTC m=+1013.704976113" watchObservedRunningTime="2025-10-02 07:33:12.693046945 +0000 UTC m=+1013.724993232" Oct 02 07:33:12 crc kubenswrapper[4960]: I1002 07:33:12.721990 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d66c446d8-rxzk8" podStartSLOduration=12.7219452 podStartE2EDuration="12.7219452s" podCreationTimestamp="2025-10-02 07:33:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:12.701895481 +0000 UTC m=+1013.733841788" watchObservedRunningTime="2025-10-02 07:33:12.7219452 +0000 UTC m=+1013.753891487" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.177813 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de0b-account-create-6q6nn" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.185813 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b4f0-account-create-sq74s" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.310473 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlpwm\" (UniqueName: \"kubernetes.io/projected/f7fa3598-b2f0-4a29-9150-0b881db98bcb-kube-api-access-wlpwm\") pod \"f7fa3598-b2f0-4a29-9150-0b881db98bcb\" (UID: \"f7fa3598-b2f0-4a29-9150-0b881db98bcb\") " Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.310711 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5kws\" (UniqueName: \"kubernetes.io/projected/781f3c80-6287-411d-9c26-ab52d38c8291-kube-api-access-k5kws\") pod \"781f3c80-6287-411d-9c26-ab52d38c8291\" (UID: \"781f3c80-6287-411d-9c26-ab52d38c8291\") " Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.320025 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fa3598-b2f0-4a29-9150-0b881db98bcb-kube-api-access-wlpwm" (OuterVolumeSpecName: "kube-api-access-wlpwm") pod "f7fa3598-b2f0-4a29-9150-0b881db98bcb" (UID: "f7fa3598-b2f0-4a29-9150-0b881db98bcb"). InnerVolumeSpecName "kube-api-access-wlpwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.321354 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/781f3c80-6287-411d-9c26-ab52d38c8291-kube-api-access-k5kws" (OuterVolumeSpecName: "kube-api-access-k5kws") pod "781f3c80-6287-411d-9c26-ab52d38c8291" (UID: "781f3c80-6287-411d-9c26-ab52d38c8291"). InnerVolumeSpecName "kube-api-access-k5kws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.413325 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlpwm\" (UniqueName: \"kubernetes.io/projected/f7fa3598-b2f0-4a29-9150-0b881db98bcb-kube-api-access-wlpwm\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.413377 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5kws\" (UniqueName: \"kubernetes.io/projected/781f3c80-6287-411d-9c26-ab52d38c8291-kube-api-access-k5kws\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.684592 4960 generic.go:334] "Generic (PLEG): container finished" podID="8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" containerID="a495c74eb425cc66ed07cdc3565ac92fb0743250ceb3c81f3a4f528f0697997a" exitCode=0 Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.684669 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sl4rw" event={"ID":"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216","Type":"ContainerDied","Data":"a495c74eb425cc66ed07cdc3565ac92fb0743250ceb3c81f3a4f528f0697997a"} Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.686684 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-de0b-account-create-6q6nn" event={"ID":"781f3c80-6287-411d-9c26-ab52d38c8291","Type":"ContainerDied","Data":"2dcc735891037adb25eb8458251ad3475df64e645214af36c94b6f023d73b623"} Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.686723 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dcc735891037adb25eb8458251ad3475df64e645214af36c94b6f023d73b623" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.686759 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de0b-account-create-6q6nn" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.690389 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f9f4c9c8-9kk6b" event={"ID":"12495902-a857-4ea0-b5bc-178803f33c12","Type":"ContainerStarted","Data":"577ac4501e9a0ce39b630127d7651cd4d3c5b26e15121bfe0333a8b403d603b5"} Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.694681 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d66c446d8-rxzk8" event={"ID":"35082c42-a42f-4eb7-a12c-91c74a6715f4","Type":"ContainerStarted","Data":"5fbc2cfdb74bd98a2adbb2461eb07e836f20c1fedaf9e3215cada27c19d42ba1"} Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.697728 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b4f0-account-create-sq74s" Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.699777 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b4f0-account-create-sq74s" event={"ID":"f7fa3598-b2f0-4a29-9150-0b881db98bcb","Type":"ContainerDied","Data":"856c1da669a21f3d26e9c78326b80309541b60c7512b1c466bc919aa1f7a0bea"} Oct 02 07:33:13 crc kubenswrapper[4960]: I1002 07:33:13.699826 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="856c1da669a21f3d26e9c78326b80309541b60c7512b1c466bc919aa1f7a0bea" Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.159379 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a856-account-create-h8wpl" Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.349171 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6dfq\" (UniqueName: \"kubernetes.io/projected/5eb7e0fd-eb02-4f1a-af2f-d61e09104130-kube-api-access-k6dfq\") pod \"5eb7e0fd-eb02-4f1a-af2f-d61e09104130\" (UID: \"5eb7e0fd-eb02-4f1a-af2f-d61e09104130\") " Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.368528 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-59867977-hxr9v" Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.413945 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eb7e0fd-eb02-4f1a-af2f-d61e09104130-kube-api-access-k6dfq" (OuterVolumeSpecName: "kube-api-access-k6dfq") pod "5eb7e0fd-eb02-4f1a-af2f-d61e09104130" (UID: "5eb7e0fd-eb02-4f1a-af2f-d61e09104130"). InnerVolumeSpecName "kube-api-access-k6dfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.466199 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6dfq\" (UniqueName: \"kubernetes.io/projected/5eb7e0fd-eb02-4f1a-af2f-d61e09104130-kube-api-access-k6dfq\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.711644 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a856-account-create-h8wpl" Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.713184 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a856-account-create-h8wpl" event={"ID":"5eb7e0fd-eb02-4f1a-af2f-d61e09104130","Type":"ContainerDied","Data":"3cbdb4183fb1808f15e9a04cb2f69aff0c4c629ad82d161628bc4fecf6b6f6f1"} Oct 02 07:33:14 crc kubenswrapper[4960]: I1002 07:33:14.713256 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cbdb4183fb1808f15e9a04cb2f69aff0c4c629ad82d161628bc4fecf6b6f6f1" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.121663 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sl4rw" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.180750 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-config-data\") pod \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.181119 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-scripts\") pod \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.181218 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8spbc\" (UniqueName: \"kubernetes.io/projected/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-kube-api-access-8spbc\") pod \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.181282 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-logs\") pod \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.181332 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-combined-ca-bundle\") pod \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\" (UID: \"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216\") " Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.186303 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-logs" (OuterVolumeSpecName: "logs") pod "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" (UID: "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.187358 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-kube-api-access-8spbc" (OuterVolumeSpecName: "kube-api-access-8spbc") pod "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" (UID: "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216"). InnerVolumeSpecName "kube-api-access-8spbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.193129 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-scripts" (OuterVolumeSpecName: "scripts") pod "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" (UID: "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.219129 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" (UID: "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.223866 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-config-data" (OuterVolumeSpecName: "config-data") pod "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" (UID: "8478ccab-58bf-45e2-9f7a-c5a4b0bcc216"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.283507 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.283543 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.283554 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8spbc\" (UniqueName: \"kubernetes.io/projected/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-kube-api-access-8spbc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.283567 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.283576 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.726911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sl4rw" event={"ID":"8478ccab-58bf-45e2-9f7a-c5a4b0bcc216","Type":"ContainerDied","Data":"d5c6b8aa307cc639b36c4b7616c31292cbae6f58156cd70752f66751d7b97583"} Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.727592 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5c6b8aa307cc639b36c4b7616c31292cbae6f58156cd70752f66751d7b97583" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.726959 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sl4rw" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.843500 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-64f4f7488b-q47mf"] Oct 02 07:33:15 crc kubenswrapper[4960]: E1002 07:33:15.844199 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb7e0fd-eb02-4f1a-af2f-d61e09104130" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844224 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb7e0fd-eb02-4f1a-af2f-d61e09104130" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: E1002 07:33:15.844233 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="init" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844241 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="init" Oct 02 07:33:15 crc kubenswrapper[4960]: E1002 07:33:15.844260 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" containerName="placement-db-sync" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844269 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" containerName="placement-db-sync" Oct 02 07:33:15 crc kubenswrapper[4960]: E1002 07:33:15.844290 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="dnsmasq-dns" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844296 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="dnsmasq-dns" Oct 02 07:33:15 crc kubenswrapper[4960]: E1002 07:33:15.844312 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="781f3c80-6287-411d-9c26-ab52d38c8291" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844320 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="781f3c80-6287-411d-9c26-ab52d38c8291" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: E1002 07:33:15.844329 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fa3598-b2f0-4a29-9150-0b881db98bcb" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844336 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fa3598-b2f0-4a29-9150-0b881db98bcb" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844535 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a13a796e-8fc2-463c-9587-8fc19145de29" containerName="dnsmasq-dns" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844552 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7fa3598-b2f0-4a29-9150-0b881db98bcb" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844576 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" containerName="placement-db-sync" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844591 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="781f3c80-6287-411d-9c26-ab52d38c8291" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.844609 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eb7e0fd-eb02-4f1a-af2f-d61e09104130" containerName="mariadb-account-create" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.859496 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.866847 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.866893 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.866943 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.867192 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wd5ms" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.867862 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.875926 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-64f4f7488b-q47mf"] Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.910523 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmhkm\" (UniqueName: \"kubernetes.io/projected/9c89938d-910f-4d87-85a5-22bfc7cc24e9-kube-api-access-cmhkm\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.910868 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-config-data\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.910949 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-combined-ca-bundle\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.911127 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c89938d-910f-4d87-85a5-22bfc7cc24e9-logs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.911207 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-scripts\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.911297 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-public-tls-certs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:15 crc kubenswrapper[4960]: I1002 07:33:15.911380 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-internal-tls-certs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.013650 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmhkm\" (UniqueName: \"kubernetes.io/projected/9c89938d-910f-4d87-85a5-22bfc7cc24e9-kube-api-access-cmhkm\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.013741 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-config-data\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.013763 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-combined-ca-bundle\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.013804 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c89938d-910f-4d87-85a5-22bfc7cc24e9-logs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.013827 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-scripts\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.013866 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-public-tls-certs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.013888 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-internal-tls-certs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.014709 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c89938d-910f-4d87-85a5-22bfc7cc24e9-logs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.021559 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-config-data\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.022769 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-internal-tls-certs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.036935 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-scripts\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.037549 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-public-tls-certs\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.040716 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c89938d-910f-4d87-85a5-22bfc7cc24e9-combined-ca-bundle\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.048684 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmhkm\" (UniqueName: \"kubernetes.io/projected/9c89938d-910f-4d87-85a5-22bfc7cc24e9-kube-api-access-cmhkm\") pod \"placement-64f4f7488b-q47mf\" (UID: \"9c89938d-910f-4d87-85a5-22bfc7cc24e9\") " pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.226072 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.740615 4960 generic.go:334] "Generic (PLEG): container finished" podID="14972b11-92c6-4bea-bf0e-f926c18c198f" containerID="95932ea4bc707c32ada6f68dd3182b3dd177a3928dce54959f6136978a031a5f" exitCode=0 Oct 02 07:33:16 crc kubenswrapper[4960]: I1002 07:33:16.740686 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xjxcq" event={"ID":"14972b11-92c6-4bea-bf0e-f926c18c198f","Type":"ContainerDied","Data":"95932ea4bc707c32ada6f68dd3182b3dd177a3928dce54959f6136978a031a5f"} Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.099607 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-x2tp5"] Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.102355 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.104961 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.105234 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-cs8v7" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.120173 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-x2tp5"] Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.178774 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-db-sync-config-data\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.178861 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwdrv\" (UniqueName: \"kubernetes.io/projected/0f2b624f-e186-4f60-898b-bce9781f911a-kube-api-access-rwdrv\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.178893 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-combined-ca-bundle\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.222717 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-965xq"] Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.224112 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.227937 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.231812 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.232022 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xrg77" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.256344 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-965xq"] Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282419 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-combined-ca-bundle\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282484 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-scripts\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282547 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-db-sync-config-data\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282599 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-db-sync-config-data\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282625 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq8pt\" (UniqueName: \"kubernetes.io/projected/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-kube-api-access-tq8pt\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282653 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwdrv\" (UniqueName: \"kubernetes.io/projected/0f2b624f-e186-4f60-898b-bce9781f911a-kube-api-access-rwdrv\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282676 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-config-data\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282702 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-combined-ca-bundle\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.282738 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-etc-machine-id\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.293949 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-db-sync-config-data\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.303898 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-combined-ca-bundle\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.309675 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwdrv\" (UniqueName: \"kubernetes.io/projected/0f2b624f-e186-4f60-898b-bce9781f911a-kube-api-access-rwdrv\") pod \"barbican-db-sync-x2tp5\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.385970 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-config-data\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.386127 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-etc-machine-id\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.386227 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-combined-ca-bundle\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.386269 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-scripts\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.386261 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-etc-machine-id\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.386327 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-db-sync-config-data\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.386394 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq8pt\" (UniqueName: \"kubernetes.io/projected/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-kube-api-access-tq8pt\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.394461 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-db-sync-config-data\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.395332 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-combined-ca-bundle\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.395396 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-scripts\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.397462 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-config-data\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.419627 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-fjshw"] Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.421241 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.421425 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq8pt\" (UniqueName: \"kubernetes.io/projected/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-kube-api-access-tq8pt\") pod \"cinder-db-sync-965xq\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.423869 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.424079 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.424260 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jw8jn" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.427441 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.434615 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fjshw"] Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.489929 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-combined-ca-bundle\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.490022 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rd5k\" (UniqueName: \"kubernetes.io/projected/b896e9b5-c55b-4c73-b6d3-36aad785df9f-kube-api-access-4rd5k\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.490089 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-config\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.542345 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-965xq" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.591145 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-config\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.591292 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-combined-ca-bundle\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.591382 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rd5k\" (UniqueName: \"kubernetes.io/projected/b896e9b5-c55b-4c73-b6d3-36aad785df9f-kube-api-access-4rd5k\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.615048 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-combined-ca-bundle\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.619776 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-config\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.625912 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rd5k\" (UniqueName: \"kubernetes.io/projected/b896e9b5-c55b-4c73-b6d3-36aad785df9f-kube-api-access-4rd5k\") pod \"neutron-db-sync-fjshw\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:19 crc kubenswrapper[4960]: I1002 07:33:19.791242 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.721330 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.808193 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xjxcq" event={"ID":"14972b11-92c6-4bea-bf0e-f926c18c198f","Type":"ContainerDied","Data":"c1304378d67c72a51dc58fcdef28b6eb7fcf61469dcb3d43deac6d88d9363630"} Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.808934 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1304378d67c72a51dc58fcdef28b6eb7fcf61469dcb3d43deac6d88d9363630" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.808511 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xjxcq" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.819191 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-config-data\") pod \"14972b11-92c6-4bea-bf0e-f926c18c198f\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.819256 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-scripts\") pod \"14972b11-92c6-4bea-bf0e-f926c18c198f\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.819320 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlf9s\" (UniqueName: \"kubernetes.io/projected/14972b11-92c6-4bea-bf0e-f926c18c198f-kube-api-access-xlf9s\") pod \"14972b11-92c6-4bea-bf0e-f926c18c198f\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.819462 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-fernet-keys\") pod \"14972b11-92c6-4bea-bf0e-f926c18c198f\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.819589 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-credential-keys\") pod \"14972b11-92c6-4bea-bf0e-f926c18c198f\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.819658 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-combined-ca-bundle\") pod \"14972b11-92c6-4bea-bf0e-f926c18c198f\" (UID: \"14972b11-92c6-4bea-bf0e-f926c18c198f\") " Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.825348 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "14972b11-92c6-4bea-bf0e-f926c18c198f" (UID: "14972b11-92c6-4bea-bf0e-f926c18c198f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.828515 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14972b11-92c6-4bea-bf0e-f926c18c198f-kube-api-access-xlf9s" (OuterVolumeSpecName: "kube-api-access-xlf9s") pod "14972b11-92c6-4bea-bf0e-f926c18c198f" (UID: "14972b11-92c6-4bea-bf0e-f926c18c198f"). InnerVolumeSpecName "kube-api-access-xlf9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.830269 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-scripts" (OuterVolumeSpecName: "scripts") pod "14972b11-92c6-4bea-bf0e-f926c18c198f" (UID: "14972b11-92c6-4bea-bf0e-f926c18c198f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.856691 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-config-data" (OuterVolumeSpecName: "config-data") pod "14972b11-92c6-4bea-bf0e-f926c18c198f" (UID: "14972b11-92c6-4bea-bf0e-f926c18c198f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.857541 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "14972b11-92c6-4bea-bf0e-f926c18c198f" (UID: "14972b11-92c6-4bea-bf0e-f926c18c198f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.861468 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14972b11-92c6-4bea-bf0e-f926c18c198f" (UID: "14972b11-92c6-4bea-bf0e-f926c18c198f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.923704 4960 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.923742 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.923753 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.923783 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.923795 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlf9s\" (UniqueName: \"kubernetes.io/projected/14972b11-92c6-4bea-bf0e-f926c18c198f-kube-api-access-xlf9s\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:20 crc kubenswrapper[4960]: I1002 07:33:20.923809 4960 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14972b11-92c6-4bea-bf0e-f926c18c198f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.124475 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.125637 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.224572 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.225703 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.249866 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-64f4f7488b-q47mf"] Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.257116 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-965xq"] Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.345598 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-x2tp5"] Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.408197 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fjshw"] Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.838721 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34621790-60f4-4239-a57d-941591637b6e","Type":"ContainerStarted","Data":"c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730"} Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.864218 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64f4f7488b-q47mf" event={"ID":"9c89938d-910f-4d87-85a5-22bfc7cc24e9","Type":"ContainerStarted","Data":"f63f11be0f5b472ffc61047f41e9dc601ae475dad38ae63cba23563365e39bdb"} Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.864307 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64f4f7488b-q47mf" event={"ID":"9c89938d-910f-4d87-85a5-22bfc7cc24e9","Type":"ContainerStarted","Data":"5afab17967ab5b8f04fb3b0f6942295e96b1ef61100b1e3e2541c42b256efb2b"} Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.866569 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-567f974799-pm8bc"] Oct 02 07:33:21 crc kubenswrapper[4960]: E1002 07:33:21.867045 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14972b11-92c6-4bea-bf0e-f926c18c198f" containerName="keystone-bootstrap" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.867066 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="14972b11-92c6-4bea-bf0e-f926c18c198f" containerName="keystone-bootstrap" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.867245 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="14972b11-92c6-4bea-bf0e-f926c18c198f" containerName="keystone-bootstrap" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.867855 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.870677 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.874666 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.874674 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.874915 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.875083 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.875157 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mplxm" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.894651 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-567f974799-pm8bc"] Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.903595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-965xq" event={"ID":"f6f22ba4-0d4e-4012-a82d-b6a98d10b015","Type":"ContainerStarted","Data":"56a3a7d88d490ddfe5ec12146f0a9be6a3e7c434098a8d900e25c4dddd25e7a0"} Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.905816 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x2tp5" event={"ID":"0f2b624f-e186-4f60-898b-bce9781f911a","Type":"ContainerStarted","Data":"b15a5490606d2da1ae649bd1970bef09599ff169ea5a62b8f4ef9ddfbd454325"} Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.909294 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fjshw" event={"ID":"b896e9b5-c55b-4c73-b6d3-36aad785df9f","Type":"ContainerStarted","Data":"51b4bafe895ab6b053d85cf47465eb80c368cba4018eb7c13cd4de98511ea9e4"} Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.909404 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fjshw" event={"ID":"b896e9b5-c55b-4c73-b6d3-36aad785df9f","Type":"ContainerStarted","Data":"668d8a433b180700bb15cfe0173f61a0619ff90e1ec1c9ed8046d6a5118173a7"} Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.964866 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-fjshw" podStartSLOduration=2.964841432 podStartE2EDuration="2.964841432s" podCreationTimestamp="2025-10-02 07:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:21.938517132 +0000 UTC m=+1022.970463429" watchObservedRunningTime="2025-10-02 07:33:21.964841432 +0000 UTC m=+1022.996787719" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.965780 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-fernet-keys\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.966114 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-internal-tls-certs\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.966222 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-config-data\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.966311 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-credential-keys\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.972706 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-public-tls-certs\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.974736 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cljlr\" (UniqueName: \"kubernetes.io/projected/ced7f231-e268-4f66-87ae-b32b809bfa6f-kube-api-access-cljlr\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.981139 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-scripts\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:21 crc kubenswrapper[4960]: I1002 07:33:21.981765 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-combined-ca-bundle\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.083596 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-scripts\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.083663 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-combined-ca-bundle\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.083730 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-fernet-keys\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.083998 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-internal-tls-certs\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.084022 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-config-data\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.084044 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-credential-keys\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.084080 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-public-tls-certs\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.084136 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cljlr\" (UniqueName: \"kubernetes.io/projected/ced7f231-e268-4f66-87ae-b32b809bfa6f-kube-api-access-cljlr\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.099565 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-scripts\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.099638 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-internal-tls-certs\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.101436 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-fernet-keys\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.101670 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-config-data\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.103352 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-public-tls-certs\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.106673 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-credential-keys\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.106923 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced7f231-e268-4f66-87ae-b32b809bfa6f-combined-ca-bundle\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.107152 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cljlr\" (UniqueName: \"kubernetes.io/projected/ced7f231-e268-4f66-87ae-b32b809bfa6f-kube-api-access-cljlr\") pod \"keystone-567f974799-pm8bc\" (UID: \"ced7f231-e268-4f66-87ae-b32b809bfa6f\") " pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.205126 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.820088 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-567f974799-pm8bc"] Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.922872 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64f4f7488b-q47mf" event={"ID":"9c89938d-910f-4d87-85a5-22bfc7cc24e9","Type":"ContainerStarted","Data":"0b79090fb8cd8c6602861915a5f02bb8631774dac5513d94130f602e6d367ca5"} Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.924052 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.924078 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.940138 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-567f974799-pm8bc" event={"ID":"ced7f231-e268-4f66-87ae-b32b809bfa6f","Type":"ContainerStarted","Data":"d3e774e64f2c8cce3624c6b3392d085fe9e681736eb5fae1067f88a4c0d152ca"} Oct 02 07:33:22 crc kubenswrapper[4960]: I1002 07:33:22.966967 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-64f4f7488b-q47mf" podStartSLOduration=7.966939462 podStartE2EDuration="7.966939462s" podCreationTimestamp="2025-10-02 07:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:22.952432573 +0000 UTC m=+1023.984378860" watchObservedRunningTime="2025-10-02 07:33:22.966939462 +0000 UTC m=+1023.998885749" Oct 02 07:33:23 crc kubenswrapper[4960]: I1002 07:33:23.952558 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-567f974799-pm8bc" event={"ID":"ced7f231-e268-4f66-87ae-b32b809bfa6f","Type":"ContainerStarted","Data":"07967a4fdccafc8698bd5faae4cc9d9083022ed5177c7137724b30c9b3634f64"} Oct 02 07:33:23 crc kubenswrapper[4960]: I1002 07:33:23.986597 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-567f974799-pm8bc" podStartSLOduration=2.986542868 podStartE2EDuration="2.986542868s" podCreationTimestamp="2025-10-02 07:33:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:23.969803915 +0000 UTC m=+1025.001750202" watchObservedRunningTime="2025-10-02 07:33:23.986542868 +0000 UTC m=+1025.018489155" Oct 02 07:33:24 crc kubenswrapper[4960]: I1002 07:33:24.973478 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:31 crc kubenswrapper[4960]: I1002 07:33:31.127779 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 07:33:31 crc kubenswrapper[4960]: I1002 07:33:31.223136 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-d66c446d8-rxzk8" podUID="35082c42-a42f-4eb7-a12c-91c74a6715f4" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Oct 02 07:33:40 crc kubenswrapper[4960]: E1002 07:33:40.041738 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:dec0378fb73f66dd39f6d7f7bdc6020d9bdf4c57c201eac3c63d7e97667627fe" Oct 02 07:33:40 crc kubenswrapper[4960]: E1002 07:33:40.043205 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:dec0378fb73f66dd39f6d7f7bdc6020d9bdf4c57c201eac3c63d7e97667627fe,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tq8pt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-965xq_openstack(f6f22ba4-0d4e-4012-a82d-b6a98d10b015): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:33:40 crc kubenswrapper[4960]: E1002 07:33:40.045230 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-965xq" podUID="f6f22ba4-0d4e-4012-a82d-b6a98d10b015" Oct 02 07:33:40 crc kubenswrapper[4960]: I1002 07:33:40.138589 4960 generic.go:334] "Generic (PLEG): container finished" podID="b896e9b5-c55b-4c73-b6d3-36aad785df9f" containerID="51b4bafe895ab6b053d85cf47465eb80c368cba4018eb7c13cd4de98511ea9e4" exitCode=0 Oct 02 07:33:40 crc kubenswrapper[4960]: I1002 07:33:40.138677 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fjshw" event={"ID":"b896e9b5-c55b-4c73-b6d3-36aad785df9f","Type":"ContainerDied","Data":"51b4bafe895ab6b053d85cf47465eb80c368cba4018eb7c13cd4de98511ea9e4"} Oct 02 07:33:40 crc kubenswrapper[4960]: E1002 07:33:40.142444 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:dec0378fb73f66dd39f6d7f7bdc6020d9bdf4c57c201eac3c63d7e97667627fe\\\"\"" pod="openstack/cinder-db-sync-965xq" podUID="f6f22ba4-0d4e-4012-a82d-b6a98d10b015" Oct 02 07:33:40 crc kubenswrapper[4960]: E1002 07:33:40.541659 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 02 07:33:40 crc kubenswrapper[4960]: E1002 07:33:40.542488 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lxqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(34621790-60f4-4239-a57d-941591637b6e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:33:40 crc kubenswrapper[4960]: E1002 07:33:40.543737 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="34621790-60f4-4239-a57d-941591637b6e" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.153843 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x2tp5" event={"ID":"0f2b624f-e186-4f60-898b-bce9781f911a","Type":"ContainerStarted","Data":"6c3207f8696d3f361033f4517227510e73b39529b4b8a2cdd4a0eb78c4149878"} Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.154027 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="ceilometer-notification-agent" containerID="cri-o://5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0" gracePeriod=30 Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.154171 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="sg-core" containerID="cri-o://c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730" gracePeriod=30 Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.197809 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-x2tp5" podStartSLOduration=3.538802737 podStartE2EDuration="22.197774005s" podCreationTimestamp="2025-10-02 07:33:19 +0000 UTC" firstStartedPulling="2025-10-02 07:33:21.34932673 +0000 UTC m=+1022.381273017" lastFinishedPulling="2025-10-02 07:33:40.008297988 +0000 UTC m=+1041.040244285" observedRunningTime="2025-10-02 07:33:41.181704287 +0000 UTC m=+1042.213650614" watchObservedRunningTime="2025-10-02 07:33:41.197774005 +0000 UTC m=+1042.229720302" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.762671 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.871860 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rd5k\" (UniqueName: \"kubernetes.io/projected/b896e9b5-c55b-4c73-b6d3-36aad785df9f-kube-api-access-4rd5k\") pod \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.872401 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-combined-ca-bundle\") pod \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.872615 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-config\") pod \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\" (UID: \"b896e9b5-c55b-4c73-b6d3-36aad785df9f\") " Oct 02 07:33:41 crc kubenswrapper[4960]: E1002 07:33:41.877173 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1f5833e_dad0_46e3_b460_b6be789f3c62.slice/crio-ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1f5833e_dad0_46e3_b460_b6be789f3c62.slice/crio-conmon-ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1f5833e_dad0_46e3_b460_b6be789f3c62.slice/crio-conmon-fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34621790_60f4_4239_a57d_941591637b6e.slice/crio-conmon-c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1f5833e_dad0_46e3_b460_b6be789f3c62.slice/crio-fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb896e9b5_c55b_4c73_b6d3_36aad785df9f.slice/crio-51b4bafe895ab6b053d85cf47465eb80c368cba4018eb7c13cd4de98511ea9e4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb896e9b5_c55b_4c73_b6d3_36aad785df9f.slice/crio-conmon-51b4bafe895ab6b053d85cf47465eb80c368cba4018eb7c13cd4de98511ea9e4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34621790_60f4_4239_a57d_941591637b6e.slice/crio-c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.882292 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b896e9b5-c55b-4c73-b6d3-36aad785df9f-kube-api-access-4rd5k" (OuterVolumeSpecName: "kube-api-access-4rd5k") pod "b896e9b5-c55b-4c73-b6d3-36aad785df9f" (UID: "b896e9b5-c55b-4c73-b6d3-36aad785df9f"). InnerVolumeSpecName "kube-api-access-4rd5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.916199 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b896e9b5-c55b-4c73-b6d3-36aad785df9f" (UID: "b896e9b5-c55b-4c73-b6d3-36aad785df9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.932671 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-config" (OuterVolumeSpecName: "config") pod "b896e9b5-c55b-4c73-b6d3-36aad785df9f" (UID: "b896e9b5-c55b-4c73-b6d3-36aad785df9f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.974962 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.975027 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b896e9b5-c55b-4c73-b6d3-36aad785df9f-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.975043 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rd5k\" (UniqueName: \"kubernetes.io/projected/b896e9b5-c55b-4c73-b6d3-36aad785df9f-kube-api-access-4rd5k\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:41 crc kubenswrapper[4960]: I1002 07:33:41.996570 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59867977-hxr9v" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.077031 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f5kc\" (UniqueName: \"kubernetes.io/projected/a1f5833e-dad0-46e3-b460-b6be789f3c62-kube-api-access-4f5kc\") pod \"a1f5833e-dad0-46e3-b460-b6be789f3c62\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.077107 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1f5833e-dad0-46e3-b460-b6be789f3c62-horizon-secret-key\") pod \"a1f5833e-dad0-46e3-b460-b6be789f3c62\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.077168 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-config-data\") pod \"a1f5833e-dad0-46e3-b460-b6be789f3c62\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.077299 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f5833e-dad0-46e3-b460-b6be789f3c62-logs\") pod \"a1f5833e-dad0-46e3-b460-b6be789f3c62\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.077332 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-scripts\") pod \"a1f5833e-dad0-46e3-b460-b6be789f3c62\" (UID: \"a1f5833e-dad0-46e3-b460-b6be789f3c62\") " Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.081089 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f5833e-dad0-46e3-b460-b6be789f3c62-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a1f5833e-dad0-46e3-b460-b6be789f3c62" (UID: "a1f5833e-dad0-46e3-b460-b6be789f3c62"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.081550 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1f5833e-dad0-46e3-b460-b6be789f3c62-logs" (OuterVolumeSpecName: "logs") pod "a1f5833e-dad0-46e3-b460-b6be789f3c62" (UID: "a1f5833e-dad0-46e3-b460-b6be789f3c62"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.087254 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1f5833e-dad0-46e3-b460-b6be789f3c62-kube-api-access-4f5kc" (OuterVolumeSpecName: "kube-api-access-4f5kc") pod "a1f5833e-dad0-46e3-b460-b6be789f3c62" (UID: "a1f5833e-dad0-46e3-b460-b6be789f3c62"). InnerVolumeSpecName "kube-api-access-4f5kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.104426 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-config-data" (OuterVolumeSpecName: "config-data") pod "a1f5833e-dad0-46e3-b460-b6be789f3c62" (UID: "a1f5833e-dad0-46e3-b460-b6be789f3c62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.112469 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-scripts" (OuterVolumeSpecName: "scripts") pod "a1f5833e-dad0-46e3-b460-b6be789f3c62" (UID: "a1f5833e-dad0-46e3-b460-b6be789f3c62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.168247 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34621790-60f4-4239-a57d-941591637b6e","Type":"ContainerDied","Data":"c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730"} Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.168102 4960 generic.go:334] "Generic (PLEG): container finished" podID="34621790-60f4-4239-a57d-941591637b6e" containerID="c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730" exitCode=2 Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.171750 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fjshw" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.171789 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fjshw" event={"ID":"b896e9b5-c55b-4c73-b6d3-36aad785df9f","Type":"ContainerDied","Data":"668d8a433b180700bb15cfe0173f61a0619ff90e1ec1c9ed8046d6a5118173a7"} Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.171834 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="668d8a433b180700bb15cfe0173f61a0619ff90e1ec1c9ed8046d6a5118173a7" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.174835 4960 generic.go:334] "Generic (PLEG): container finished" podID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerID="fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2" exitCode=137 Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.174870 4960 generic.go:334] "Generic (PLEG): container finished" podID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerID="ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc" exitCode=137 Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.174958 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59867977-hxr9v" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.175006 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59867977-hxr9v" event={"ID":"a1f5833e-dad0-46e3-b460-b6be789f3c62","Type":"ContainerDied","Data":"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2"} Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.175096 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59867977-hxr9v" event={"ID":"a1f5833e-dad0-46e3-b460-b6be789f3c62","Type":"ContainerDied","Data":"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc"} Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.175119 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59867977-hxr9v" event={"ID":"a1f5833e-dad0-46e3-b460-b6be789f3c62","Type":"ContainerDied","Data":"360fe37038f042a1e95da8bd4c8166ffd9535b5364a2708ea07b72210d4100fd"} Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.175149 4960 scope.go:117] "RemoveContainer" containerID="fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.179033 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f5kc\" (UniqueName: \"kubernetes.io/projected/a1f5833e-dad0-46e3-b460-b6be789f3c62-kube-api-access-4f5kc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.179059 4960 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1f5833e-dad0-46e3-b460-b6be789f3c62-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.179070 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.179080 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f5833e-dad0-46e3-b460-b6be789f3c62-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.179090 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1f5833e-dad0-46e3-b460-b6be789f3c62-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.237404 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59867977-hxr9v"] Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.280257 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-59867977-hxr9v"] Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.387711 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" path="/var/lib/kubelet/pods/a1f5833e-dad0-46e3-b460-b6be789f3c62/volumes" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.443745 4960 scope.go:117] "RemoveContainer" containerID="ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.498811 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58f9f745b5-2vzqx"] Oct 02 07:33:42 crc kubenswrapper[4960]: E1002 07:33:42.499298 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.499312 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon" Oct 02 07:33:42 crc kubenswrapper[4960]: E1002 07:33:42.499346 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon-log" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.499352 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon-log" Oct 02 07:33:42 crc kubenswrapper[4960]: E1002 07:33:42.499376 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b896e9b5-c55b-4c73-b6d3-36aad785df9f" containerName="neutron-db-sync" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.499382 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b896e9b5-c55b-4c73-b6d3-36aad785df9f" containerName="neutron-db-sync" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.499576 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon-log" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.499591 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b896e9b5-c55b-4c73-b6d3-36aad785df9f" containerName="neutron-db-sync" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.499603 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1f5833e-dad0-46e3-b460-b6be789f3c62" containerName="horizon" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.500682 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.501655 4960 scope.go:117] "RemoveContainer" containerID="fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2" Oct 02 07:33:42 crc kubenswrapper[4960]: E1002 07:33:42.507830 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2\": container with ID starting with fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2 not found: ID does not exist" containerID="fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.507885 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2"} err="failed to get container status \"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2\": rpc error: code = NotFound desc = could not find container \"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2\": container with ID starting with fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2 not found: ID does not exist" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.507919 4960 scope.go:117] "RemoveContainer" containerID="ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc" Oct 02 07:33:42 crc kubenswrapper[4960]: E1002 07:33:42.508234 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc\": container with ID starting with ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc not found: ID does not exist" containerID="ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.508251 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc"} err="failed to get container status \"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc\": rpc error: code = NotFound desc = could not find container \"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc\": container with ID starting with ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc not found: ID does not exist" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.508266 4960 scope.go:117] "RemoveContainer" containerID="fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.508528 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2"} err="failed to get container status \"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2\": rpc error: code = NotFound desc = could not find container \"fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2\": container with ID starting with fb16cc9ed91865fda7fd0a3cf494b9a0b26db998006b6620cbe0136bc7fbb3f2 not found: ID does not exist" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.508569 4960 scope.go:117] "RemoveContainer" containerID="ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.516403 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc"} err="failed to get container status \"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc\": rpc error: code = NotFound desc = could not find container \"ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc\": container with ID starting with ff56ff3b9113840f01fddb3c91d3473dfe4cc0b99844f4df81a2f61eb19de2bc not found: ID does not exist" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.523779 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58f9f745b5-2vzqx"] Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.602725 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgc8x\" (UniqueName: \"kubernetes.io/projected/99535389-fde0-45d9-9c25-b27541aee476-kube-api-access-mgc8x\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.603066 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-dns-svc\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.603395 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-nb\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.603600 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-sb\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.603778 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-config\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.611508 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6ff76df558-2w955"] Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.617913 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.622671 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.622760 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.622798 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jw8jn" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.623050 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.636461 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ff76df558-2w955"] Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.706311 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-nb\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.706404 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-sb\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.706459 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-config\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.706492 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgc8x\" (UniqueName: \"kubernetes.io/projected/99535389-fde0-45d9-9c25-b27541aee476-kube-api-access-mgc8x\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.706522 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-dns-svc\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.707518 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-nb\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.712475 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-dns-svc\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.713557 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-config\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.719122 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-sb\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.737201 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgc8x\" (UniqueName: \"kubernetes.io/projected/99535389-fde0-45d9-9c25-b27541aee476-kube-api-access-mgc8x\") pod \"dnsmasq-dns-58f9f745b5-2vzqx\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.808881 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-combined-ca-bundle\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.808950 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-httpd-config\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.809003 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-config\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.809120 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-ovndb-tls-certs\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.809145 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmrw8\" (UniqueName: \"kubernetes.io/projected/19b6c4e4-068d-4969-a8b3-95d907bde1c5-kube-api-access-zmrw8\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.844466 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.911291 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmrw8\" (UniqueName: \"kubernetes.io/projected/19b6c4e4-068d-4969-a8b3-95d907bde1c5-kube-api-access-zmrw8\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.911459 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-combined-ca-bundle\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.911494 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-httpd-config\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.911516 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-config\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.911583 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-ovndb-tls-certs\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.919672 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-config\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.921704 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-httpd-config\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.925674 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-ovndb-tls-certs\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.925947 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-combined-ca-bundle\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.934148 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmrw8\" (UniqueName: \"kubernetes.io/projected/19b6c4e4-068d-4969-a8b3-95d907bde1c5-kube-api-access-zmrw8\") pod \"neutron-6ff76df558-2w955\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:42 crc kubenswrapper[4960]: I1002 07:33:42.937813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:43 crc kubenswrapper[4960]: I1002 07:33:43.422046 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58f9f745b5-2vzqx"] Oct 02 07:33:43 crc kubenswrapper[4960]: W1002 07:33:43.440191 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99535389_fde0_45d9_9c25_b27541aee476.slice/crio-bc317d0fee983a5b3cf3b0e03ba9a9dbc8451d6e23fcda21c45a2ccdb24d6df8 WatchSource:0}: Error finding container bc317d0fee983a5b3cf3b0e03ba9a9dbc8451d6e23fcda21c45a2ccdb24d6df8: Status 404 returned error can't find the container with id bc317d0fee983a5b3cf3b0e03ba9a9dbc8451d6e23fcda21c45a2ccdb24d6df8 Oct 02 07:33:43 crc kubenswrapper[4960]: I1002 07:33:43.822037 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ff76df558-2w955"] Oct 02 07:33:43 crc kubenswrapper[4960]: W1002 07:33:43.833131 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19b6c4e4_068d_4969_a8b3_95d907bde1c5.slice/crio-57042676f06a48c73dcb789fde695ef5f7d352e31e62f782c3b084c640dcc8ce WatchSource:0}: Error finding container 57042676f06a48c73dcb789fde695ef5f7d352e31e62f782c3b084c640dcc8ce: Status 404 returned error can't find the container with id 57042676f06a48c73dcb789fde695ef5f7d352e31e62f782c3b084c640dcc8ce Oct 02 07:33:43 crc kubenswrapper[4960]: I1002 07:33:43.847347 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:43 crc kubenswrapper[4960]: I1002 07:33:43.932746 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:33:44 crc kubenswrapper[4960]: I1002 07:33:44.198083 4960 generic.go:334] "Generic (PLEG): container finished" podID="99535389-fde0-45d9-9c25-b27541aee476" containerID="55c08f9673f5a7042a01846edb00e33e42e1212c3a17be5a057ddecec3c06813" exitCode=0 Oct 02 07:33:44 crc kubenswrapper[4960]: I1002 07:33:44.198208 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" event={"ID":"99535389-fde0-45d9-9c25-b27541aee476","Type":"ContainerDied","Data":"55c08f9673f5a7042a01846edb00e33e42e1212c3a17be5a057ddecec3c06813"} Oct 02 07:33:44 crc kubenswrapper[4960]: I1002 07:33:44.198578 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" event={"ID":"99535389-fde0-45d9-9c25-b27541aee476","Type":"ContainerStarted","Data":"bc317d0fee983a5b3cf3b0e03ba9a9dbc8451d6e23fcda21c45a2ccdb24d6df8"} Oct 02 07:33:44 crc kubenswrapper[4960]: I1002 07:33:44.201767 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ff76df558-2w955" event={"ID":"19b6c4e4-068d-4969-a8b3-95d907bde1c5","Type":"ContainerStarted","Data":"665ef4457fcb60e709404f75b3aa015ad8af4cd231ac3754feb33527c9f3946e"} Oct 02 07:33:44 crc kubenswrapper[4960]: I1002 07:33:44.201799 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ff76df558-2w955" event={"ID":"19b6c4e4-068d-4969-a8b3-95d907bde1c5","Type":"ContainerStarted","Data":"57042676f06a48c73dcb789fde695ef5f7d352e31e62f782c3b084c640dcc8ce"} Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.017959 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d8fb8597c-r7rqs"] Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.021689 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.025642 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.030609 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.038422 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d8fb8597c-r7rqs"] Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.165674 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-config\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.165738 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-ovndb-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.165766 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-combined-ca-bundle\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.165851 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-public-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.165890 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-httpd-config\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.165913 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwzm2\" (UniqueName: \"kubernetes.io/projected/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-kube-api-access-dwzm2\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.165959 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-internal-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.247120 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.247248 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" event={"ID":"99535389-fde0-45d9-9c25-b27541aee476","Type":"ContainerStarted","Data":"96775d1e40bd302bb44878b282dda9f788b2ac2cf82dee3b4f75097260a8b265"} Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.247573 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.250888 4960 generic.go:334] "Generic (PLEG): container finished" podID="34621790-60f4-4239-a57d-941591637b6e" containerID="5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0" exitCode=0 Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.250957 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34621790-60f4-4239-a57d-941591637b6e","Type":"ContainerDied","Data":"5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0"} Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.251076 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34621790-60f4-4239-a57d-941591637b6e","Type":"ContainerDied","Data":"b7bdf6901b4749974ad0f83e4bb54c4b671eea632a1c0780e8d6598ca273b14a"} Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.251101 4960 scope.go:117] "RemoveContainer" containerID="c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.258332 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ff76df558-2w955" event={"ID":"19b6c4e4-068d-4969-a8b3-95d907bde1c5","Type":"ContainerStarted","Data":"a33ccfa61b073f58fe5079e93ef170a2a45cb9a82220d4f291da1ff7065f246f"} Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.258820 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.261765 4960 generic.go:334] "Generic (PLEG): container finished" podID="0f2b624f-e186-4f60-898b-bce9781f911a" containerID="6c3207f8696d3f361033f4517227510e73b39529b4b8a2cdd4a0eb78c4149878" exitCode=0 Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.261794 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x2tp5" event={"ID":"0f2b624f-e186-4f60-898b-bce9781f911a","Type":"ContainerDied","Data":"6c3207f8696d3f361033f4517227510e73b39529b4b8a2cdd4a0eb78c4149878"} Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.271176 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-public-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.271250 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-httpd-config\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.271279 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwzm2\" (UniqueName: \"kubernetes.io/projected/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-kube-api-access-dwzm2\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.271344 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-internal-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.271380 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-config\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.271405 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-ovndb-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.271425 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-combined-ca-bundle\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.276054 4960 scope.go:117] "RemoveContainer" containerID="5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.289127 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-combined-ca-bundle\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.289371 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-ovndb-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.289742 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-httpd-config\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.305210 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-config\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.305222 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-internal-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.308214 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-public-tls-certs\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.308512 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwzm2\" (UniqueName: \"kubernetes.io/projected/4e13c181-da1b-4da1-9ed7-daa1d05f0f5f-kube-api-access-dwzm2\") pod \"neutron-5d8fb8597c-r7rqs\" (UID: \"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f\") " pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.309139 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" podStartSLOduration=3.309113311 podStartE2EDuration="3.309113311s" podCreationTimestamp="2025-10-02 07:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:45.269418273 +0000 UTC m=+1046.301364560" watchObservedRunningTime="2025-10-02 07:33:45.309113311 +0000 UTC m=+1046.341059598" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.340823 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.347193 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6ff76df558-2w955" podStartSLOduration=3.347177024 podStartE2EDuration="3.347177024s" podCreationTimestamp="2025-10-02 07:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:45.335821021 +0000 UTC m=+1046.367767308" watchObservedRunningTime="2025-10-02 07:33:45.347177024 +0000 UTC m=+1046.379123311" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.357672 4960 scope.go:117] "RemoveContainer" containerID="c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730" Oct 02 07:33:45 crc kubenswrapper[4960]: E1002 07:33:45.360747 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730\": container with ID starting with c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730 not found: ID does not exist" containerID="c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.360823 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730"} err="failed to get container status \"c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730\": rpc error: code = NotFound desc = could not find container \"c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730\": container with ID starting with c52c19c99feb9359e11ba5362ed1bc88e717e586682dbe0a85353b0e2c27d730 not found: ID does not exist" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.360868 4960 scope.go:117] "RemoveContainer" containerID="5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0" Oct 02 07:33:45 crc kubenswrapper[4960]: E1002 07:33:45.361337 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0\": container with ID starting with 5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0 not found: ID does not exist" containerID="5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.361389 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0"} err="failed to get container status \"5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0\": rpc error: code = NotFound desc = could not find container \"5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0\": container with ID starting with 5d76398df324f69e979aa79c38af697ae215292a3f9ad158e7c546b4ce7c24d0 not found: ID does not exist" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.379726 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-log-httpd\") pod \"34621790-60f4-4239-a57d-941591637b6e\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.379809 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-sg-core-conf-yaml\") pod \"34621790-60f4-4239-a57d-941591637b6e\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.379955 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-combined-ca-bundle\") pod \"34621790-60f4-4239-a57d-941591637b6e\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.380011 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lxqm\" (UniqueName: \"kubernetes.io/projected/34621790-60f4-4239-a57d-941591637b6e-kube-api-access-9lxqm\") pod \"34621790-60f4-4239-a57d-941591637b6e\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.380061 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-scripts\") pod \"34621790-60f4-4239-a57d-941591637b6e\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.380106 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-run-httpd\") pod \"34621790-60f4-4239-a57d-941591637b6e\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.380170 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-config-data\") pod \"34621790-60f4-4239-a57d-941591637b6e\" (UID: \"34621790-60f4-4239-a57d-941591637b6e\") " Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.380991 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "34621790-60f4-4239-a57d-941591637b6e" (UID: "34621790-60f4-4239-a57d-941591637b6e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.384044 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "34621790-60f4-4239-a57d-941591637b6e" (UID: "34621790-60f4-4239-a57d-941591637b6e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.388141 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34621790-60f4-4239-a57d-941591637b6e-kube-api-access-9lxqm" (OuterVolumeSpecName: "kube-api-access-9lxqm") pod "34621790-60f4-4239-a57d-941591637b6e" (UID: "34621790-60f4-4239-a57d-941591637b6e"). InnerVolumeSpecName "kube-api-access-9lxqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.388253 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-scripts" (OuterVolumeSpecName: "scripts") pod "34621790-60f4-4239-a57d-941591637b6e" (UID: "34621790-60f4-4239-a57d-941591637b6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.416138 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34621790-60f4-4239-a57d-941591637b6e" (UID: "34621790-60f4-4239-a57d-941591637b6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.425563 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "34621790-60f4-4239-a57d-941591637b6e" (UID: "34621790-60f4-4239-a57d-941591637b6e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.437143 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-config-data" (OuterVolumeSpecName: "config-data") pod "34621790-60f4-4239-a57d-941591637b6e" (UID: "34621790-60f4-4239-a57d-941591637b6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.489885 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.489919 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.489928 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.489940 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.489950 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lxqm\" (UniqueName: \"kubernetes.io/projected/34621790-60f4-4239-a57d-941591637b6e-kube-api-access-9lxqm\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.489959 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34621790-60f4-4239-a57d-941591637b6e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.489967 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34621790-60f4-4239-a57d-941591637b6e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.900279 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-d66c446d8-rxzk8" Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.975337 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85f9f4c9c8-9kk6b"] Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.975595 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon-log" containerID="cri-o://18bae37076d32c0f6e54e8ed4eb2151d38367a0781878ae2c2c5ffe57acd4e5f" gracePeriod=30 Oct 02 07:33:45 crc kubenswrapper[4960]: I1002 07:33:45.976052 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" containerID="cri-o://577ac4501e9a0ce39b630127d7651cd4d3c5b26e15121bfe0333a8b403d603b5" gracePeriod=30 Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.000307 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d8fb8597c-r7rqs"] Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.001557 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.271916 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d8fb8597c-r7rqs" event={"ID":"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f","Type":"ContainerStarted","Data":"1760ef9bfa462ce77c97164b557f0ad359db36ab1a8a04eeb3eba0570b9d3dc8"} Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.273456 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.384070 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.399495 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.421075 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:33:46 crc kubenswrapper[4960]: E1002 07:33:46.421637 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="ceilometer-notification-agent" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.421660 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="ceilometer-notification-agent" Oct 02 07:33:46 crc kubenswrapper[4960]: E1002 07:33:46.421720 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="sg-core" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.421729 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="sg-core" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.421936 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="ceilometer-notification-agent" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.421965 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="34621790-60f4-4239-a57d-941591637b6e" containerName="sg-core" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.425863 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.432236 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.432513 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.443626 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.528110 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj5kd\" (UniqueName: \"kubernetes.io/projected/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-kube-api-access-mj5kd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.528157 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-run-httpd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.528188 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-config-data\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.528208 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.528247 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.528288 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-scripts\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.528338 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-log-httpd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.634099 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-log-httpd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.634175 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj5kd\" (UniqueName: \"kubernetes.io/projected/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-kube-api-access-mj5kd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.634205 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-run-httpd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.634224 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-config-data\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.634245 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.634276 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.634324 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-scripts\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.635408 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-log-httpd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.639440 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-run-httpd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.646690 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.647996 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.653994 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-config-data\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.655878 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-scripts\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.692566 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj5kd\" (UniqueName: \"kubernetes.io/projected/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-kube-api-access-mj5kd\") pod \"ceilometer-0\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " pod="openstack/ceilometer-0" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.788600 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.838509 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-combined-ca-bundle\") pod \"0f2b624f-e186-4f60-898b-bce9781f911a\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.838691 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-db-sync-config-data\") pod \"0f2b624f-e186-4f60-898b-bce9781f911a\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.838729 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwdrv\" (UniqueName: \"kubernetes.io/projected/0f2b624f-e186-4f60-898b-bce9781f911a-kube-api-access-rwdrv\") pod \"0f2b624f-e186-4f60-898b-bce9781f911a\" (UID: \"0f2b624f-e186-4f60-898b-bce9781f911a\") " Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.845275 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0f2b624f-e186-4f60-898b-bce9781f911a" (UID: "0f2b624f-e186-4f60-898b-bce9781f911a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.855234 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2b624f-e186-4f60-898b-bce9781f911a-kube-api-access-rwdrv" (OuterVolumeSpecName: "kube-api-access-rwdrv") pod "0f2b624f-e186-4f60-898b-bce9781f911a" (UID: "0f2b624f-e186-4f60-898b-bce9781f911a"). InnerVolumeSpecName "kube-api-access-rwdrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.875389 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f2b624f-e186-4f60-898b-bce9781f911a" (UID: "0f2b624f-e186-4f60-898b-bce9781f911a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.940568 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.940607 4960 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f2b624f-e186-4f60-898b-bce9781f911a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.940617 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwdrv\" (UniqueName: \"kubernetes.io/projected/0f2b624f-e186-4f60-898b-bce9781f911a-kube-api-access-rwdrv\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:46 crc kubenswrapper[4960]: I1002 07:33:46.960782 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.295187 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x2tp5" event={"ID":"0f2b624f-e186-4f60-898b-bce9781f911a","Type":"ContainerDied","Data":"b15a5490606d2da1ae649bd1970bef09599ff169ea5a62b8f4ef9ddfbd454325"} Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.295586 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b15a5490606d2da1ae649bd1970bef09599ff169ea5a62b8f4ef9ddfbd454325" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.295665 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x2tp5" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.304178 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d8fb8597c-r7rqs" event={"ID":"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f","Type":"ContainerStarted","Data":"14cf4d6c2e18dfebb3c1d1d4f363d8fa6a15698463abe707a177b7724f93ba60"} Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.304256 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d8fb8597c-r7rqs" event={"ID":"4e13c181-da1b-4da1-9ed7-daa1d05f0f5f","Type":"ContainerStarted","Data":"d7571d4eb31a1d182a6890126af48423d12ee43e1ef2391c457000864c6adab1"} Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.304280 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.333672 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d8fb8597c-r7rqs" podStartSLOduration=3.333609884 podStartE2EDuration="3.333609884s" podCreationTimestamp="2025-10-02 07:33:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:47.330832873 +0000 UTC m=+1048.362779160" watchObservedRunningTime="2025-10-02 07:33:47.333609884 +0000 UTC m=+1048.365556171" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.531667 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.610759 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5bd958db77-xxp6q"] Oct 02 07:33:47 crc kubenswrapper[4960]: E1002 07:33:47.611522 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2b624f-e186-4f60-898b-bce9781f911a" containerName="barbican-db-sync" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.611547 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2b624f-e186-4f60-898b-bce9781f911a" containerName="barbican-db-sync" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.611715 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2b624f-e186-4f60-898b-bce9781f911a" containerName="barbican-db-sync" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.612867 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.620984 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.621126 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.621184 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-cs8v7" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.661201 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-84cf7698d6-jhmj6"] Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.663370 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.666189 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-config-data-custom\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.666349 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnl5w\" (UniqueName: \"kubernetes.io/projected/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-kube-api-access-qnl5w\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.666433 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-logs\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.666488 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-config-data\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.666529 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-combined-ca-bundle\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.673018 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.701531 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5bd958db77-xxp6q"] Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.718313 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84cf7698d6-jhmj6"] Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768254 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-config-data-custom\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768376 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc171097-1fed-426b-9aa6-1c5c905fa5c6-logs\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768413 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnl5w\" (UniqueName: \"kubernetes.io/projected/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-kube-api-access-qnl5w\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768546 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6z7h\" (UniqueName: \"kubernetes.io/projected/cc171097-1fed-426b-9aa6-1c5c905fa5c6-kube-api-access-k6z7h\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768599 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-logs\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768634 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-config-data\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768674 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-combined-ca-bundle\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768696 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-config-data\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768728 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-combined-ca-bundle\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.768842 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-config-data-custom\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.771879 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-logs\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.778401 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-config-data\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.793993 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-config-data-custom\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.796791 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-combined-ca-bundle\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.825688 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnl5w\" (UniqueName: \"kubernetes.io/projected/2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7-kube-api-access-qnl5w\") pod \"barbican-keystone-listener-84cf7698d6-jhmj6\" (UID: \"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7\") " pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.848941 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58f9f745b5-2vzqx"] Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.849543 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" podUID="99535389-fde0-45d9-9c25-b27541aee476" containerName="dnsmasq-dns" containerID="cri-o://96775d1e40bd302bb44878b282dda9f788b2ac2cf82dee3b4f75097260a8b265" gracePeriod=10 Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.872830 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-combined-ca-bundle\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.873048 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-config-data\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.873422 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-config-data-custom\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.874156 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc171097-1fed-426b-9aa6-1c5c905fa5c6-logs\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.874192 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6z7h\" (UniqueName: \"kubernetes.io/projected/cc171097-1fed-426b-9aa6-1c5c905fa5c6-kube-api-access-k6z7h\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.874774 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc171097-1fed-426b-9aa6-1c5c905fa5c6-logs\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.896878 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-combined-ca-bundle\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.908602 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-config-data\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.911156 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7854d55f69-7h9wq"] Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.911685 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc171097-1fed-426b-9aa6-1c5c905fa5c6-config-data-custom\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.911930 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6z7h\" (UniqueName: \"kubernetes.io/projected/cc171097-1fed-426b-9aa6-1c5c905fa5c6-kube-api-access-k6z7h\") pod \"barbican-worker-5bd958db77-xxp6q\" (UID: \"cc171097-1fed-426b-9aa6-1c5c905fa5c6\") " pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.924861 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.943812 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7854d55f69-7h9wq"] Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.952571 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5bd958db77-xxp6q" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.976317 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-config\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.976372 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffp6c\" (UniqueName: \"kubernetes.io/projected/f8d36716-f1a9-4ce1-8692-65638a5c8884-kube-api-access-ffp6c\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.976467 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-sb\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.976494 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-dns-svc\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:47 crc kubenswrapper[4960]: I1002 07:33:47.976541 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-nb\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.041457 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.080665 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-sb\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.080761 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-dns-svc\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.080942 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-nb\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.081013 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-config\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.081065 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffp6c\" (UniqueName: \"kubernetes.io/projected/f8d36716-f1a9-4ce1-8692-65638a5c8884-kube-api-access-ffp6c\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.082080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-sb\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.082584 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-dns-svc\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.082716 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-config\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.083281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-nb\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.101425 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-76d58bd87b-72mnv"] Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.103103 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.104769 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffp6c\" (UniqueName: \"kubernetes.io/projected/f8d36716-f1a9-4ce1-8692-65638a5c8884-kube-api-access-ffp6c\") pod \"dnsmasq-dns-7854d55f69-7h9wq\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.115611 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.116680 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76d58bd87b-72mnv"] Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.184354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wswm5\" (UniqueName: \"kubernetes.io/projected/8e92f003-5585-452b-9d1e-3915e1ec6f2a-kube-api-access-wswm5\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.184868 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data-custom\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.184900 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.184940 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e92f003-5585-452b-9d1e-3915e1ec6f2a-logs\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.185113 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-combined-ca-bundle\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.279748 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.287381 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e92f003-5585-452b-9d1e-3915e1ec6f2a-logs\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.287455 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-combined-ca-bundle\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.287528 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wswm5\" (UniqueName: \"kubernetes.io/projected/8e92f003-5585-452b-9d1e-3915e1ec6f2a-kube-api-access-wswm5\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.287599 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data-custom\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.287632 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.287874 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e92f003-5585-452b-9d1e-3915e1ec6f2a-logs\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.294438 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-combined-ca-bundle\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.300547 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data-custom\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.303593 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.322091 4960 generic.go:334] "Generic (PLEG): container finished" podID="99535389-fde0-45d9-9c25-b27541aee476" containerID="96775d1e40bd302bb44878b282dda9f788b2ac2cf82dee3b4f75097260a8b265" exitCode=0 Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.322428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" event={"ID":"99535389-fde0-45d9-9c25-b27541aee476","Type":"ContainerDied","Data":"96775d1e40bd302bb44878b282dda9f788b2ac2cf82dee3b4f75097260a8b265"} Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.324588 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wswm5\" (UniqueName: \"kubernetes.io/projected/8e92f003-5585-452b-9d1e-3915e1ec6f2a-kube-api-access-wswm5\") pod \"barbican-api-76d58bd87b-72mnv\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.328153 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerStarted","Data":"42dbc9a9b9bac9d8dd7aa5b717f3c17cf173c8b9a173820dc5fdc375dd38e4e1"} Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.340647 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34621790-60f4-4239-a57d-941591637b6e" path="/var/lib/kubelet/pods/34621790-60f4-4239-a57d-941591637b6e/volumes" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.436772 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.808591 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:48 crc kubenswrapper[4960]: I1002 07:33:48.978036 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-64f4f7488b-q47mf" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.274012 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.333588 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7854d55f69-7h9wq"] Oct 02 07:33:49 crc kubenswrapper[4960]: W1002 07:33:49.340197 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8d36716_f1a9_4ce1_8692_65638a5c8884.slice/crio-8011f7be15fa11b0d8c7d3a81d1953e8bbd1d2f94f0699cdd459b0dcef9763e1 WatchSource:0}: Error finding container 8011f7be15fa11b0d8c7d3a81d1953e8bbd1d2f94f0699cdd459b0dcef9763e1: Status 404 returned error can't find the container with id 8011f7be15fa11b0d8c7d3a81d1953e8bbd1d2f94f0699cdd459b0dcef9763e1 Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.352048 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerStarted","Data":"6357085e3742e99be8025a127b2e686d8425f689d1795173cfe1ef5bceb967f2"} Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.359618 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" event={"ID":"99535389-fde0-45d9-9c25-b27541aee476","Type":"ContainerDied","Data":"bc317d0fee983a5b3cf3b0e03ba9a9dbc8451d6e23fcda21c45a2ccdb24d6df8"} Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.359694 4960 scope.go:117] "RemoveContainer" containerID="96775d1e40bd302bb44878b282dda9f788b2ac2cf82dee3b4f75097260a8b265" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.359943 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f9f745b5-2vzqx" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.384587 4960 scope.go:117] "RemoveContainer" containerID="55c08f9673f5a7042a01846edb00e33e42e1212c3a17be5a057ddecec3c06813" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.431714 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-nb\") pod \"99535389-fde0-45d9-9c25-b27541aee476\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.431775 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-dns-svc\") pod \"99535389-fde0-45d9-9c25-b27541aee476\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.431904 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-sb\") pod \"99535389-fde0-45d9-9c25-b27541aee476\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.431943 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-config\") pod \"99535389-fde0-45d9-9c25-b27541aee476\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.432082 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgc8x\" (UniqueName: \"kubernetes.io/projected/99535389-fde0-45d9-9c25-b27541aee476-kube-api-access-mgc8x\") pod \"99535389-fde0-45d9-9c25-b27541aee476\" (UID: \"99535389-fde0-45d9-9c25-b27541aee476\") " Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.456909 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99535389-fde0-45d9-9c25-b27541aee476-kube-api-access-mgc8x" (OuterVolumeSpecName: "kube-api-access-mgc8x") pod "99535389-fde0-45d9-9c25-b27541aee476" (UID: "99535389-fde0-45d9-9c25-b27541aee476"). InnerVolumeSpecName "kube-api-access-mgc8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.499660 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99535389-fde0-45d9-9c25-b27541aee476" (UID: "99535389-fde0-45d9-9c25-b27541aee476"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.522374 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-config" (OuterVolumeSpecName: "config") pod "99535389-fde0-45d9-9c25-b27541aee476" (UID: "99535389-fde0-45d9-9c25-b27541aee476"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.535638 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.535673 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.535684 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgc8x\" (UniqueName: \"kubernetes.io/projected/99535389-fde0-45d9-9c25-b27541aee476-kube-api-access-mgc8x\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.551132 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "99535389-fde0-45d9-9c25-b27541aee476" (UID: "99535389-fde0-45d9-9c25-b27541aee476"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.595368 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "99535389-fde0-45d9-9c25-b27541aee476" (UID: "99535389-fde0-45d9-9c25-b27541aee476"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.632999 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76d58bd87b-72mnv"] Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.638143 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.638179 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99535389-fde0-45d9-9c25-b27541aee476-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.661646 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5bd958db77-xxp6q"] Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.712339 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84cf7698d6-jhmj6"] Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.815030 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58f9f745b5-2vzqx"] Oct 02 07:33:49 crc kubenswrapper[4960]: I1002 07:33:49.827326 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58f9f745b5-2vzqx"] Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.342626 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99535389-fde0-45d9-9c25-b27541aee476" path="/var/lib/kubelet/pods/99535389-fde0-45d9-9c25-b27541aee476/volumes" Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.417689 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bd958db77-xxp6q" event={"ID":"cc171097-1fed-426b-9aa6-1c5c905fa5c6","Type":"ContainerStarted","Data":"1cfb4f7c3c02a20dc7bbac98f9acd790e2752efd7b2972a7035e8cc07fc9c3c0"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.440329 4960 generic.go:334] "Generic (PLEG): container finished" podID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerID="7409f5d4cb1f48fb62b1a5358be6470cc3b12ddfa36d61bd06a75f778e5b9b69" exitCode=0 Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.440482 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" event={"ID":"f8d36716-f1a9-4ce1-8692-65638a5c8884","Type":"ContainerDied","Data":"7409f5d4cb1f48fb62b1a5358be6470cc3b12ddfa36d61bd06a75f778e5b9b69"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.440532 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" event={"ID":"f8d36716-f1a9-4ce1-8692-65638a5c8884","Type":"ContainerStarted","Data":"8011f7be15fa11b0d8c7d3a81d1953e8bbd1d2f94f0699cdd459b0dcef9763e1"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.451619 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d58bd87b-72mnv" event={"ID":"8e92f003-5585-452b-9d1e-3915e1ec6f2a","Type":"ContainerStarted","Data":"49b602fa88e894c950210ca54cd12180f1f11b32cfd783e22adda8ef2b9bf414"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.451676 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d58bd87b-72mnv" event={"ID":"8e92f003-5585-452b-9d1e-3915e1ec6f2a","Type":"ContainerStarted","Data":"681877e74714d583554fe8d657317e977c5760cde38447998da93ccb223a0623"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.451688 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d58bd87b-72mnv" event={"ID":"8e92f003-5585-452b-9d1e-3915e1ec6f2a","Type":"ContainerStarted","Data":"7d09bfbbf318e6776baddda85a65dd748258ab7de9a51650681ca2bb171dff55"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.452810 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.452845 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.463777 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" event={"ID":"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7","Type":"ContainerStarted","Data":"bd137b2e3a9402b62ce7e90e57fb6ece10fd83728cb708bc411a5fb8430565c7"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.474637 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerStarted","Data":"df79319fe38ef1ee5a6208330b070a59bb459cb8673857a3d4c25b3d82c6170d"} Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.514289 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:34494->10.217.0.141:8443: read: connection reset by peer" Oct 02 07:33:50 crc kubenswrapper[4960]: I1002 07:33:50.528662 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-76d58bd87b-72mnv" podStartSLOduration=2.5286387809999997 podStartE2EDuration="2.528638781s" podCreationTimestamp="2025-10-02 07:33:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:50.527021804 +0000 UTC m=+1051.558968091" watchObservedRunningTime="2025-10-02 07:33:50.528638781 +0000 UTC m=+1051.560585068" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.125448 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.288889 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-69bd96fd88-9gstt"] Oct 02 07:33:51 crc kubenswrapper[4960]: E1002 07:33:51.289417 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99535389-fde0-45d9-9c25-b27541aee476" containerName="dnsmasq-dns" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.289436 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="99535389-fde0-45d9-9c25-b27541aee476" containerName="dnsmasq-dns" Oct 02 07:33:51 crc kubenswrapper[4960]: E1002 07:33:51.289464 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99535389-fde0-45d9-9c25-b27541aee476" containerName="init" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.289472 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="99535389-fde0-45d9-9c25-b27541aee476" containerName="init" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.289631 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="99535389-fde0-45d9-9c25-b27541aee476" containerName="dnsmasq-dns" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.292949 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.296804 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.303026 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.306228 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-69bd96fd88-9gstt"] Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.373868 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-internal-tls-certs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.373941 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6fd5\" (UniqueName: \"kubernetes.io/projected/76d6cdce-7551-483b-859c-678805e24895-kube-api-access-v6fd5\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.373963 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d6cdce-7551-483b-859c-678805e24895-logs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.374020 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-config-data-custom\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.374086 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-combined-ca-bundle\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.374112 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-config-data\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.374158 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-public-tls-certs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.476042 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-combined-ca-bundle\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.476118 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-config-data\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.476186 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-public-tls-certs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.476298 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-internal-tls-certs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.476332 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6fd5\" (UniqueName: \"kubernetes.io/projected/76d6cdce-7551-483b-859c-678805e24895-kube-api-access-v6fd5\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.476353 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d6cdce-7551-483b-859c-678805e24895-logs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.476385 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-config-data-custom\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.479303 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76d6cdce-7551-483b-859c-678805e24895-logs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.483515 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-internal-tls-certs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.488366 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-config-data-custom\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.488581 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-config-data\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.489013 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-public-tls-certs\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.492418 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d6cdce-7551-483b-859c-678805e24895-combined-ca-bundle\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.503532 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6fd5\" (UniqueName: \"kubernetes.io/projected/76d6cdce-7551-483b-859c-678805e24895-kube-api-access-v6fd5\") pod \"barbican-api-69bd96fd88-9gstt\" (UID: \"76d6cdce-7551-483b-859c-678805e24895\") " pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.505456 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" event={"ID":"f8d36716-f1a9-4ce1-8692-65638a5c8884","Type":"ContainerStarted","Data":"d9dbb6f4486feaa14174f1bbc5786cd99bc3f06a015aa5db3f9de9c459d6e088"} Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.505737 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.520423 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerStarted","Data":"199fc2d4ad72ce0da2819fc6e6feaddc57685dc00eae6739440498a42c69e902"} Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.524022 4960 generic.go:334] "Generic (PLEG): container finished" podID="12495902-a857-4ea0-b5bc-178803f33c12" containerID="577ac4501e9a0ce39b630127d7651cd4d3c5b26e15121bfe0333a8b403d603b5" exitCode=0 Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.525040 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f9f4c9c8-9kk6b" event={"ID":"12495902-a857-4ea0-b5bc-178803f33c12","Type":"ContainerDied","Data":"577ac4501e9a0ce39b630127d7651cd4d3c5b26e15121bfe0333a8b403d603b5"} Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.531725 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" podStartSLOduration=4.531709735 podStartE2EDuration="4.531709735s" podCreationTimestamp="2025-10-02 07:33:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:51.527606255 +0000 UTC m=+1052.559552542" watchObservedRunningTime="2025-10-02 07:33:51.531709735 +0000 UTC m=+1052.563656022" Oct 02 07:33:51 crc kubenswrapper[4960]: I1002 07:33:51.617274 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:52 crc kubenswrapper[4960]: I1002 07:33:52.957381 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-69bd96fd88-9gstt"] Oct 02 07:33:53 crc kubenswrapper[4960]: W1002 07:33:53.010693 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76d6cdce_7551_483b_859c_678805e24895.slice/crio-2aa29ea3d139e19a012b2b6879cf5f4acd0c83c022aaf2901a50c50a67a894c8 WatchSource:0}: Error finding container 2aa29ea3d139e19a012b2b6879cf5f4acd0c83c022aaf2901a50c50a67a894c8: Status 404 returned error can't find the container with id 2aa29ea3d139e19a012b2b6879cf5f4acd0c83c022aaf2901a50c50a67a894c8 Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.561413 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bd958db77-xxp6q" event={"ID":"cc171097-1fed-426b-9aa6-1c5c905fa5c6","Type":"ContainerStarted","Data":"14d4b63a30ba502b3e9b5f93351a03c06af432f8f5bb829b7a105f33e134066d"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.561466 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bd958db77-xxp6q" event={"ID":"cc171097-1fed-426b-9aa6-1c5c905fa5c6","Type":"ContainerStarted","Data":"c4024e2f49da42bfe338dfbd1f718900c5c85b0e7ae39bfc595f4b8f2dea522b"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.564279 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" event={"ID":"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7","Type":"ContainerStarted","Data":"5c361699cd6fdc143a2a52f2e76dce52e5fa08541185dff10eed3bfbca52eacc"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.564308 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" event={"ID":"2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7","Type":"ContainerStarted","Data":"ac55d9de4c37bb9b41d8916277d80e0b8b94c2acfe2a128c52ba69795f19ee67"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.567360 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerStarted","Data":"9e96d525fba70d1ccbdc935eea63997f47269c6b8a4a58beb2523f88d1d262bd"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.567878 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.571064 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69bd96fd88-9gstt" event={"ID":"76d6cdce-7551-483b-859c-678805e24895","Type":"ContainerStarted","Data":"e5556f42bf5a2f2c32cf092a4ed33820d61248d110fff5f276d973ee23c92d42"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.571090 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69bd96fd88-9gstt" event={"ID":"76d6cdce-7551-483b-859c-678805e24895","Type":"ContainerStarted","Data":"5297a57e19b8a1277ca9b302391b6e441ebc91724d75b9e93e2c6dcc643e8e36"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.571123 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69bd96fd88-9gstt" event={"ID":"76d6cdce-7551-483b-859c-678805e24895","Type":"ContainerStarted","Data":"2aa29ea3d139e19a012b2b6879cf5f4acd0c83c022aaf2901a50c50a67a894c8"} Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.571276 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.571298 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.594406 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5bd958db77-xxp6q" podStartSLOduration=3.901608484 podStartE2EDuration="6.594379203s" podCreationTimestamp="2025-10-02 07:33:47 +0000 UTC" firstStartedPulling="2025-10-02 07:33:49.693594075 +0000 UTC m=+1050.725540352" lastFinishedPulling="2025-10-02 07:33:52.386364784 +0000 UTC m=+1053.418311071" observedRunningTime="2025-10-02 07:33:53.585938046 +0000 UTC m=+1054.617884333" watchObservedRunningTime="2025-10-02 07:33:53.594379203 +0000 UTC m=+1054.626325490" Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.619407 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-69bd96fd88-9gstt" podStartSLOduration=2.619389103 podStartE2EDuration="2.619389103s" podCreationTimestamp="2025-10-02 07:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:53.614624933 +0000 UTC m=+1054.646571220" watchObservedRunningTime="2025-10-02 07:33:53.619389103 +0000 UTC m=+1054.651335380" Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.651578 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.82777291 podStartE2EDuration="7.651557893s" podCreationTimestamp="2025-10-02 07:33:46 +0000 UTC" firstStartedPulling="2025-10-02 07:33:47.55774526 +0000 UTC m=+1048.589691547" lastFinishedPulling="2025-10-02 07:33:52.381530243 +0000 UTC m=+1053.413476530" observedRunningTime="2025-10-02 07:33:53.640987143 +0000 UTC m=+1054.672933430" watchObservedRunningTime="2025-10-02 07:33:53.651557893 +0000 UTC m=+1054.683504180" Oct 02 07:33:53 crc kubenswrapper[4960]: I1002 07:33:53.657651 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-84cf7698d6-jhmj6" podStartSLOduration=4.022305797 podStartE2EDuration="6.657619719s" podCreationTimestamp="2025-10-02 07:33:47 +0000 UTC" firstStartedPulling="2025-10-02 07:33:49.747752536 +0000 UTC m=+1050.779698823" lastFinishedPulling="2025-10-02 07:33:52.383066448 +0000 UTC m=+1053.415012745" observedRunningTime="2025-10-02 07:33:53.656744573 +0000 UTC m=+1054.688690860" watchObservedRunningTime="2025-10-02 07:33:53.657619719 +0000 UTC m=+1054.689566006" Oct 02 07:33:54 crc kubenswrapper[4960]: I1002 07:33:54.265458 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-567f974799-pm8bc" Oct 02 07:33:55 crc kubenswrapper[4960]: I1002 07:33:55.597480 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-965xq" event={"ID":"f6f22ba4-0d4e-4012-a82d-b6a98d10b015","Type":"ContainerStarted","Data":"177b934ededc13b521097dde3f4552f990e83b874819bd756af4d397b05dec25"} Oct 02 07:33:55 crc kubenswrapper[4960]: I1002 07:33:55.622968 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-965xq" podStartSLOduration=3.128021497 podStartE2EDuration="36.622944925s" podCreationTimestamp="2025-10-02 07:33:19 +0000 UTC" firstStartedPulling="2025-10-02 07:33:21.268396822 +0000 UTC m=+1022.300343109" lastFinishedPulling="2025-10-02 07:33:54.76332023 +0000 UTC m=+1055.795266537" observedRunningTime="2025-10-02 07:33:55.619062651 +0000 UTC m=+1056.651008938" watchObservedRunningTime="2025-10-02 07:33:55.622944925 +0000 UTC m=+1056.654891202" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.017892 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.019372 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.026081 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.033615 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.033886 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-896gv" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.062192 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.125332 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34704ecb-b08d-4f67-b6ed-66141a0f6e74-combined-ca-bundle\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.125878 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/34704ecb-b08d-4f67-b6ed-66141a0f6e74-openstack-config\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.125910 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/34704ecb-b08d-4f67-b6ed-66141a0f6e74-openstack-config-secret\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.125945 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fcqf\" (UniqueName: \"kubernetes.io/projected/34704ecb-b08d-4f67-b6ed-66141a0f6e74-kube-api-access-2fcqf\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.228564 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34704ecb-b08d-4f67-b6ed-66141a0f6e74-combined-ca-bundle\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.228642 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/34704ecb-b08d-4f67-b6ed-66141a0f6e74-openstack-config\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.228668 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/34704ecb-b08d-4f67-b6ed-66141a0f6e74-openstack-config-secret\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.228700 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fcqf\" (UniqueName: \"kubernetes.io/projected/34704ecb-b08d-4f67-b6ed-66141a0f6e74-kube-api-access-2fcqf\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.230066 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/34704ecb-b08d-4f67-b6ed-66141a0f6e74-openstack-config\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.237530 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/34704ecb-b08d-4f67-b6ed-66141a0f6e74-openstack-config-secret\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.237809 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34704ecb-b08d-4f67-b6ed-66141a0f6e74-combined-ca-bundle\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.246534 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fcqf\" (UniqueName: \"kubernetes.io/projected/34704ecb-b08d-4f67-b6ed-66141a0f6e74-kube-api-access-2fcqf\") pod \"openstackclient\" (UID: \"34704ecb-b08d-4f67-b6ed-66141a0f6e74\") " pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.353024 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 07:33:56 crc kubenswrapper[4960]: I1002 07:33:56.901013 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 07:33:57 crc kubenswrapper[4960]: I1002 07:33:57.639127 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"34704ecb-b08d-4f67-b6ed-66141a0f6e74","Type":"ContainerStarted","Data":"64367196200eabaf490aaf90f78a0d85e281d2f35e331bd48670259f7f7d6ace"} Oct 02 07:33:58 crc kubenswrapper[4960]: I1002 07:33:58.283265 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:33:58 crc kubenswrapper[4960]: I1002 07:33:58.388610 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-796bc46545-gltd7"] Oct 02 07:33:58 crc kubenswrapper[4960]: I1002 07:33:58.388884 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-796bc46545-gltd7" podUID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerName="dnsmasq-dns" containerID="cri-o://ae02f767c7cd874afd138f88d701ca1a8e2cf589d376756c330d330779956b0a" gracePeriod=10 Oct 02 07:33:58 crc kubenswrapper[4960]: I1002 07:33:58.663065 4960 generic.go:334] "Generic (PLEG): container finished" podID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerID="ae02f767c7cd874afd138f88d701ca1a8e2cf589d376756c330d330779956b0a" exitCode=0 Oct 02 07:33:58 crc kubenswrapper[4960]: I1002 07:33:58.663612 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-796bc46545-gltd7" event={"ID":"0a970f1b-0b01-4b2c-821b-a2c9660e1845","Type":"ContainerDied","Data":"ae02f767c7cd874afd138f88d701ca1a8e2cf589d376756c330d330779956b0a"} Oct 02 07:33:58 crc kubenswrapper[4960]: I1002 07:33:58.983233 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.090249 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-dns-svc\") pod \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.090392 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-sb\") pod \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.090459 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-nb\") pod \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.090525 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-config\") pod \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.090616 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdcqj\" (UniqueName: \"kubernetes.io/projected/0a970f1b-0b01-4b2c-821b-a2c9660e1845-kube-api-access-qdcqj\") pod \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\" (UID: \"0a970f1b-0b01-4b2c-821b-a2c9660e1845\") " Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.102173 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a970f1b-0b01-4b2c-821b-a2c9660e1845-kube-api-access-qdcqj" (OuterVolumeSpecName: "kube-api-access-qdcqj") pod "0a970f1b-0b01-4b2c-821b-a2c9660e1845" (UID: "0a970f1b-0b01-4b2c-821b-a2c9660e1845"). InnerVolumeSpecName "kube-api-access-qdcqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.167027 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a970f1b-0b01-4b2c-821b-a2c9660e1845" (UID: "0a970f1b-0b01-4b2c-821b-a2c9660e1845"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.171247 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a970f1b-0b01-4b2c-821b-a2c9660e1845" (UID: "0a970f1b-0b01-4b2c-821b-a2c9660e1845"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.172816 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a970f1b-0b01-4b2c-821b-a2c9660e1845" (UID: "0a970f1b-0b01-4b2c-821b-a2c9660e1845"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.183096 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-config" (OuterVolumeSpecName: "config") pod "0a970f1b-0b01-4b2c-821b-a2c9660e1845" (UID: "0a970f1b-0b01-4b2c-821b-a2c9660e1845"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.192848 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdcqj\" (UniqueName: \"kubernetes.io/projected/0a970f1b-0b01-4b2c-821b-a2c9660e1845-kube-api-access-qdcqj\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.193251 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.193262 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.193272 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.193281 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a970f1b-0b01-4b2c-821b-a2c9660e1845-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.686766 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-796bc46545-gltd7" event={"ID":"0a970f1b-0b01-4b2c-821b-a2c9660e1845","Type":"ContainerDied","Data":"1cd9c40821a47d8bf6d13680ca5eb280ed6cd1ec6429e2eb48c3e58295db860d"} Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.686835 4960 scope.go:117] "RemoveContainer" containerID="ae02f767c7cd874afd138f88d701ca1a8e2cf589d376756c330d330779956b0a" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.686891 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-796bc46545-gltd7" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.714427 4960 scope.go:117] "RemoveContainer" containerID="e1f184fdd7a6b40030c1d8b5b059a1b54690326b9efee674466aa360b36aa949" Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.735065 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-796bc46545-gltd7"] Oct 02 07:33:59 crc kubenswrapper[4960]: I1002 07:33:59.743751 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-796bc46545-gltd7"] Oct 02 07:34:00 crc kubenswrapper[4960]: I1002 07:34:00.291261 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:34:00 crc kubenswrapper[4960]: I1002 07:34:00.343740 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" path="/var/lib/kubelet/pods/0a970f1b-0b01-4b2c-821b-a2c9660e1845/volumes" Oct 02 07:34:00 crc kubenswrapper[4960]: I1002 07:34:00.703325 4960 generic.go:334] "Generic (PLEG): container finished" podID="f6f22ba4-0d4e-4012-a82d-b6a98d10b015" containerID="177b934ededc13b521097dde3f4552f990e83b874819bd756af4d397b05dec25" exitCode=0 Oct 02 07:34:00 crc kubenswrapper[4960]: I1002 07:34:00.703422 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-965xq" event={"ID":"f6f22ba4-0d4e-4012-a82d-b6a98d10b015","Type":"ContainerDied","Data":"177b934ededc13b521097dde3f4552f990e83b874819bd756af4d397b05dec25"} Oct 02 07:34:00 crc kubenswrapper[4960]: I1002 07:34:00.737019 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:34:01 crc kubenswrapper[4960]: I1002 07:34:01.125702 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.130806 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-965xq" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.268210 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-etc-machine-id\") pod \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.268279 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq8pt\" (UniqueName: \"kubernetes.io/projected/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-kube-api-access-tq8pt\") pod \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.268338 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-db-sync-config-data\") pod \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.268494 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-scripts\") pod \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.268527 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-combined-ca-bundle\") pod \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.268592 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-config-data\") pod \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\" (UID: \"f6f22ba4-0d4e-4012-a82d-b6a98d10b015\") " Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.268882 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f6f22ba4-0d4e-4012-a82d-b6a98d10b015" (UID: "f6f22ba4-0d4e-4012-a82d-b6a98d10b015"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.269035 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.296213 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-scripts" (OuterVolumeSpecName: "scripts") pod "f6f22ba4-0d4e-4012-a82d-b6a98d10b015" (UID: "f6f22ba4-0d4e-4012-a82d-b6a98d10b015"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.297565 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f6f22ba4-0d4e-4012-a82d-b6a98d10b015" (UID: "f6f22ba4-0d4e-4012-a82d-b6a98d10b015"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.310203 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-kube-api-access-tq8pt" (OuterVolumeSpecName: "kube-api-access-tq8pt") pod "f6f22ba4-0d4e-4012-a82d-b6a98d10b015" (UID: "f6f22ba4-0d4e-4012-a82d-b6a98d10b015"). InnerVolumeSpecName "kube-api-access-tq8pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.320878 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6f22ba4-0d4e-4012-a82d-b6a98d10b015" (UID: "f6f22ba4-0d4e-4012-a82d-b6a98d10b015"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.372648 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq8pt\" (UniqueName: \"kubernetes.io/projected/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-kube-api-access-tq8pt\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.372704 4960 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.372718 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.372730 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.426158 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-config-data" (OuterVolumeSpecName: "config-data") pod "f6f22ba4-0d4e-4012-a82d-b6a98d10b015" (UID: "f6f22ba4-0d4e-4012-a82d-b6a98d10b015"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.475739 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f22ba4-0d4e-4012-a82d-b6a98d10b015-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.729917 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-965xq" event={"ID":"f6f22ba4-0d4e-4012-a82d-b6a98d10b015","Type":"ContainerDied","Data":"56a3a7d88d490ddfe5ec12146f0a9be6a3e7c434098a8d900e25c4dddd25e7a0"} Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.730318 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56a3a7d88d490ddfe5ec12146f0a9be6a3e7c434098a8d900e25c4dddd25e7a0" Oct 02 07:34:02 crc kubenswrapper[4960]: I1002 07:34:02.730073 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-965xq" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.126207 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d6575d66c-hm94c"] Oct 02 07:34:03 crc kubenswrapper[4960]: E1002 07:34:03.126786 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerName="dnsmasq-dns" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.126799 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerName="dnsmasq-dns" Oct 02 07:34:03 crc kubenswrapper[4960]: E1002 07:34:03.126826 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f22ba4-0d4e-4012-a82d-b6a98d10b015" containerName="cinder-db-sync" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.126832 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f22ba4-0d4e-4012-a82d-b6a98d10b015" containerName="cinder-db-sync" Oct 02 07:34:03 crc kubenswrapper[4960]: E1002 07:34:03.126853 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerName="init" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.126861 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerName="init" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.127089 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a970f1b-0b01-4b2c-821b-a2c9660e1845" containerName="dnsmasq-dns" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.127103 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f22ba4-0d4e-4012-a82d-b6a98d10b015" containerName="cinder-db-sync" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.128218 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.145655 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.155440 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.168107 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6575d66c-hm94c"] Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.168816 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.169062 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.169159 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.168837 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xrg77" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.187659 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-config\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.187832 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-dns-svc\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.187915 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.190140 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cks5f\" (UniqueName: \"kubernetes.io/projected/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-kube-api-access-cks5f\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.190277 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.196552 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.292945 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-dns-svc\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293013 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293041 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293098 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cks5f\" (UniqueName: \"kubernetes.io/projected/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-kube-api-access-cks5f\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293126 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293156 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293205 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293267 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-config\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293296 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-scripts\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293321 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.293342 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2kvp\" (UniqueName: \"kubernetes.io/projected/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-kube-api-access-j2kvp\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.294256 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-dns-svc\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.294818 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.298088 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.302210 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-config\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.309072 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.310894 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.320379 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.333103 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cks5f\" (UniqueName: \"kubernetes.io/projected/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-kube-api-access-cks5f\") pod \"dnsmasq-dns-7d6575d66c-hm94c\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.337767 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395027 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395077 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395111 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395166 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-scripts\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395262 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc6lx\" (UniqueName: \"kubernetes.io/projected/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-kube-api-access-rc6lx\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395320 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395348 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-scripts\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395409 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395430 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2kvp\" (UniqueName: \"kubernetes.io/projected/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-kube-api-access-j2kvp\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395467 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-logs\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.395500 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.396441 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.404638 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.419004 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.423349 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-scripts\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.429235 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2kvp\" (UniqueName: \"kubernetes.io/projected/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-kube-api-access-j2kvp\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.431518 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data\") pod \"cinder-scheduler-0\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.484778 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.499381 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-scripts\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.499712 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc6lx\" (UniqueName: \"kubernetes.io/projected/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-kube-api-access-rc6lx\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.499832 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.499911 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.500033 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-logs\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.500156 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.500229 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.500376 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.505245 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-logs\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.516573 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-scripts\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.517383 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.518500 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.525545 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.529638 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.530160 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc6lx\" (UniqueName: \"kubernetes.io/projected/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-kube-api-access-rc6lx\") pod \"cinder-api-0\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.689188 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.772054 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.840942 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-69bd96fd88-9gstt" Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.944240 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-76d58bd87b-72mnv"] Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.947868 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-76d58bd87b-72mnv" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api-log" containerID="cri-o://681877e74714d583554fe8d657317e977c5760cde38447998da93ccb223a0623" gracePeriod=30 Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.948402 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-76d58bd87b-72mnv" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api" containerID="cri-o://49b602fa88e894c950210ca54cd12180f1f11b32cfd783e22adda8ef2b9bf414" gracePeriod=30 Oct 02 07:34:03 crc kubenswrapper[4960]: I1002 07:34:03.963537 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76d58bd87b-72mnv" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": EOF" Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.351100 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6575d66c-hm94c"] Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.385571 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.493344 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:04 crc kubenswrapper[4960]: W1002 07:34:04.512907 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac1f219d_59a3_4c0c_ab27_03b8b3fca7e2.slice/crio-5874d4e1a4af7f11ceb99700f3f84b20efdac1f7b52098cb15c8a653b5ff832c WatchSource:0}: Error finding container 5874d4e1a4af7f11ceb99700f3f84b20efdac1f7b52098cb15c8a653b5ff832c: Status 404 returned error can't find the container with id 5874d4e1a4af7f11ceb99700f3f84b20efdac1f7b52098cb15c8a653b5ff832c Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.818192 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2","Type":"ContainerStarted","Data":"5874d4e1a4af7f11ceb99700f3f84b20efdac1f7b52098cb15c8a653b5ff832c"} Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.834245 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" event={"ID":"26d2b40e-4552-4527-b31f-8d5af0b8f7f5","Type":"ContainerStarted","Data":"792242c7aac159c2c6a9edf7a1ea5e9ea3d2279bf7d3341463ed35b1164f1782"} Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.897530 4960 generic.go:334] "Generic (PLEG): container finished" podID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerID="681877e74714d583554fe8d657317e977c5760cde38447998da93ccb223a0623" exitCode=143 Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.898049 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d58bd87b-72mnv" event={"ID":"8e92f003-5585-452b-9d1e-3915e1ec6f2a","Type":"ContainerDied","Data":"681877e74714d583554fe8d657317e977c5760cde38447998da93ccb223a0623"} Oct 02 07:34:04 crc kubenswrapper[4960]: I1002 07:34:04.918661 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31","Type":"ContainerStarted","Data":"82e7f6b11a4a91d0b4a10b333b14882778b78308dbb5110752da74f7d2fed0fe"} Oct 02 07:34:05 crc kubenswrapper[4960]: I1002 07:34:05.464726 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:05 crc kubenswrapper[4960]: I1002 07:34:05.970531 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2","Type":"ContainerStarted","Data":"247beb3ea514f9ee887529c686fcab5cf90a7fe5061f815efe48fac3e133b9a3"} Oct 02 07:34:05 crc kubenswrapper[4960]: I1002 07:34:05.986731 4960 generic.go:334] "Generic (PLEG): container finished" podID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerID="8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3" exitCode=0 Oct 02 07:34:05 crc kubenswrapper[4960]: I1002 07:34:05.987049 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" event={"ID":"26d2b40e-4552-4527-b31f-8d5af0b8f7f5","Type":"ContainerDied","Data":"8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3"} Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.088122 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-9sml2"] Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.089767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9sml2" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.095159 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9sml2"] Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.176541 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-gqr95"] Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.182007 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gqr95" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.185605 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gqr95"] Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.220071 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5km6\" (UniqueName: \"kubernetes.io/projected/b2a60afc-781b-481f-af68-4d8b9c737d29-kube-api-access-m5km6\") pod \"nova-cell0-db-create-gqr95\" (UID: \"b2a60afc-781b-481f-af68-4d8b9c737d29\") " pod="openstack/nova-cell0-db-create-gqr95" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.220130 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghlwx\" (UniqueName: \"kubernetes.io/projected/54a76d7e-e996-49f6-a3df-7da0d3005b55-kube-api-access-ghlwx\") pod \"nova-api-db-create-9sml2\" (UID: \"54a76d7e-e996-49f6-a3df-7da0d3005b55\") " pod="openstack/nova-api-db-create-9sml2" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.322167 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5km6\" (UniqueName: \"kubernetes.io/projected/b2a60afc-781b-481f-af68-4d8b9c737d29-kube-api-access-m5km6\") pod \"nova-cell0-db-create-gqr95\" (UID: \"b2a60afc-781b-481f-af68-4d8b9c737d29\") " pod="openstack/nova-cell0-db-create-gqr95" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.322640 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghlwx\" (UniqueName: \"kubernetes.io/projected/54a76d7e-e996-49f6-a3df-7da0d3005b55-kube-api-access-ghlwx\") pod \"nova-api-db-create-9sml2\" (UID: \"54a76d7e-e996-49f6-a3df-7da0d3005b55\") " pod="openstack/nova-api-db-create-9sml2" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.364001 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-tqvz5"] Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.378056 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5km6\" (UniqueName: \"kubernetes.io/projected/b2a60afc-781b-481f-af68-4d8b9c737d29-kube-api-access-m5km6\") pod \"nova-cell0-db-create-gqr95\" (UID: \"b2a60afc-781b-481f-af68-4d8b9c737d29\") " pod="openstack/nova-cell0-db-create-gqr95" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.380809 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghlwx\" (UniqueName: \"kubernetes.io/projected/54a76d7e-e996-49f6-a3df-7da0d3005b55-kube-api-access-ghlwx\") pod \"nova-api-db-create-9sml2\" (UID: \"54a76d7e-e996-49f6-a3df-7da0d3005b55\") " pod="openstack/nova-api-db-create-9sml2" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.382035 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqvz5" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.405700 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tqvz5"] Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.424784 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwgbh\" (UniqueName: \"kubernetes.io/projected/61aefc93-ba77-48df-af65-af010d8a8efa-kube-api-access-fwgbh\") pod \"nova-cell1-db-create-tqvz5\" (UID: \"61aefc93-ba77-48df-af65-af010d8a8efa\") " pod="openstack/nova-cell1-db-create-tqvz5" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.475763 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9sml2" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.503349 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gqr95" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.527341 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwgbh\" (UniqueName: \"kubernetes.io/projected/61aefc93-ba77-48df-af65-af010d8a8efa-kube-api-access-fwgbh\") pod \"nova-cell1-db-create-tqvz5\" (UID: \"61aefc93-ba77-48df-af65-af010d8a8efa\") " pod="openstack/nova-cell1-db-create-tqvz5" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.553577 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwgbh\" (UniqueName: \"kubernetes.io/projected/61aefc93-ba77-48df-af65-af010d8a8efa-kube-api-access-fwgbh\") pod \"nova-cell1-db-create-tqvz5\" (UID: \"61aefc93-ba77-48df-af65-af010d8a8efa\") " pod="openstack/nova-cell1-db-create-tqvz5" Oct 02 07:34:06 crc kubenswrapper[4960]: I1002 07:34:06.820627 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqvz5" Oct 02 07:34:07 crc kubenswrapper[4960]: I1002 07:34:07.019662 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" event={"ID":"26d2b40e-4552-4527-b31f-8d5af0b8f7f5","Type":"ContainerStarted","Data":"170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c"} Oct 02 07:34:07 crc kubenswrapper[4960]: I1002 07:34:07.020287 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:07 crc kubenswrapper[4960]: I1002 07:34:07.030519 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31","Type":"ContainerStarted","Data":"27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649"} Oct 02 07:34:07 crc kubenswrapper[4960]: I1002 07:34:07.095254 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" podStartSLOduration=4.095212857 podStartE2EDuration="4.095212857s" podCreationTimestamp="2025-10-02 07:34:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:07.048542314 +0000 UTC m=+1068.080488601" watchObservedRunningTime="2025-10-02 07:34:07.095212857 +0000 UTC m=+1068.127159134" Oct 02 07:34:07 crc kubenswrapper[4960]: I1002 07:34:07.130486 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gqr95"] Oct 02 07:34:07 crc kubenswrapper[4960]: I1002 07:34:07.149984 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9sml2"] Oct 02 07:34:07 crc kubenswrapper[4960]: I1002 07:34:07.605386 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tqvz5"] Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.050796 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2","Type":"ContainerStarted","Data":"17710c9519ac7c5849f4e05184b02837ebbaad2552ae1cb05f8014b63c3086e1"} Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.050920 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api-log" containerID="cri-o://247beb3ea514f9ee887529c686fcab5cf90a7fe5061f815efe48fac3e133b9a3" gracePeriod=30 Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.051000 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.051090 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api" containerID="cri-o://17710c9519ac7c5849f4e05184b02837ebbaad2552ae1cb05f8014b63c3086e1" gracePeriod=30 Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.054493 4960 generic.go:334] "Generic (PLEG): container finished" podID="54a76d7e-e996-49f6-a3df-7da0d3005b55" containerID="8eb073a6a9c0d3083222c384c62a22d5fbf2300ab9ee5f7e7bf0347d5a142ed9" exitCode=0 Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.054564 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9sml2" event={"ID":"54a76d7e-e996-49f6-a3df-7da0d3005b55","Type":"ContainerDied","Data":"8eb073a6a9c0d3083222c384c62a22d5fbf2300ab9ee5f7e7bf0347d5a142ed9"} Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.054615 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9sml2" event={"ID":"54a76d7e-e996-49f6-a3df-7da0d3005b55","Type":"ContainerStarted","Data":"9149b5d70188c5dbcd23e19fa93774f64f9ad7ff370776469ce7d452cb48eec1"} Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.057525 4960 generic.go:334] "Generic (PLEG): container finished" podID="b2a60afc-781b-481f-af68-4d8b9c737d29" containerID="e1c2490d5b48ed1caf344cc1538ab1d9b5e4534571f04b4ae105ca95646670ee" exitCode=0 Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.057624 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gqr95" event={"ID":"b2a60afc-781b-481f-af68-4d8b9c737d29","Type":"ContainerDied","Data":"e1c2490d5b48ed1caf344cc1538ab1d9b5e4534571f04b4ae105ca95646670ee"} Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.057663 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gqr95" event={"ID":"b2a60afc-781b-481f-af68-4d8b9c737d29","Type":"ContainerStarted","Data":"88ed6b96c0ef7df1727212dcd85ab4236e40a97975861b3aacca82d4ebf43663"} Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.072996 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.072965601 podStartE2EDuration="5.072965601s" podCreationTimestamp="2025-10-02 07:34:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:08.067919164 +0000 UTC m=+1069.099865451" watchObservedRunningTime="2025-10-02 07:34:08.072965601 +0000 UTC m=+1069.104911888" Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.087668 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31","Type":"ContainerStarted","Data":"3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc"} Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.134849 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.104549259 podStartE2EDuration="5.134827798s" podCreationTimestamp="2025-10-02 07:34:03 +0000 UTC" firstStartedPulling="2025-10-02 07:34:04.4378073 +0000 UTC m=+1065.469753587" lastFinishedPulling="2025-10-02 07:34:05.468085839 +0000 UTC m=+1066.500032126" observedRunningTime="2025-10-02 07:34:08.128804501 +0000 UTC m=+1069.160750798" watchObservedRunningTime="2025-10-02 07:34:08.134827798 +0000 UTC m=+1069.166774085" Oct 02 07:34:08 crc kubenswrapper[4960]: I1002 07:34:08.519739 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.100937 4960 generic.go:334] "Generic (PLEG): container finished" podID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerID="17710c9519ac7c5849f4e05184b02837ebbaad2552ae1cb05f8014b63c3086e1" exitCode=0 Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.100987 4960 generic.go:334] "Generic (PLEG): container finished" podID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerID="247beb3ea514f9ee887529c686fcab5cf90a7fe5061f815efe48fac3e133b9a3" exitCode=143 Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.101019 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2","Type":"ContainerDied","Data":"17710c9519ac7c5849f4e05184b02837ebbaad2552ae1cb05f8014b63c3086e1"} Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.101087 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2","Type":"ContainerDied","Data":"247beb3ea514f9ee887529c686fcab5cf90a7fe5061f815efe48fac3e133b9a3"} Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.471864 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76d58bd87b-72mnv" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:37870->10.217.0.156:9311: read: connection reset by peer" Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.471946 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76d58bd87b-72mnv" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:37856->10.217.0.156:9311: read: connection reset by peer" Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.689592 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.689940 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-central-agent" containerID="cri-o://6357085e3742e99be8025a127b2e686d8425f689d1795173cfe1ef5bceb967f2" gracePeriod=30 Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.690124 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="proxy-httpd" containerID="cri-o://9e96d525fba70d1ccbdc935eea63997f47269c6b8a4a58beb2523f88d1d262bd" gracePeriod=30 Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.690219 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-notification-agent" containerID="cri-o://df79319fe38ef1ee5a6208330b070a59bb459cb8673857a3d4c25b3d82c6170d" gracePeriod=30 Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.690276 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="sg-core" containerID="cri-o://199fc2d4ad72ce0da2819fc6e6feaddc57685dc00eae6739440498a42c69e902" gracePeriod=30 Oct 02 07:34:09 crc kubenswrapper[4960]: I1002 07:34:09.699046 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.152:3000/\": EOF" Oct 02 07:34:10 crc kubenswrapper[4960]: I1002 07:34:10.117590 4960 generic.go:334] "Generic (PLEG): container finished" podID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerID="49b602fa88e894c950210ca54cd12180f1f11b32cfd783e22adda8ef2b9bf414" exitCode=0 Oct 02 07:34:10 crc kubenswrapper[4960]: I1002 07:34:10.117685 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d58bd87b-72mnv" event={"ID":"8e92f003-5585-452b-9d1e-3915e1ec6f2a","Type":"ContainerDied","Data":"49b602fa88e894c950210ca54cd12180f1f11b32cfd783e22adda8ef2b9bf414"} Oct 02 07:34:10 crc kubenswrapper[4960]: I1002 07:34:10.122639 4960 generic.go:334] "Generic (PLEG): container finished" podID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerID="9e96d525fba70d1ccbdc935eea63997f47269c6b8a4a58beb2523f88d1d262bd" exitCode=0 Oct 02 07:34:10 crc kubenswrapper[4960]: I1002 07:34:10.122697 4960 generic.go:334] "Generic (PLEG): container finished" podID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerID="199fc2d4ad72ce0da2819fc6e6feaddc57685dc00eae6739440498a42c69e902" exitCode=2 Oct 02 07:34:10 crc kubenswrapper[4960]: I1002 07:34:10.122714 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerDied","Data":"9e96d525fba70d1ccbdc935eea63997f47269c6b8a4a58beb2523f88d1d262bd"} Oct 02 07:34:10 crc kubenswrapper[4960]: I1002 07:34:10.122794 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerDied","Data":"199fc2d4ad72ce0da2819fc6e6feaddc57685dc00eae6739440498a42c69e902"} Oct 02 07:34:11 crc kubenswrapper[4960]: I1002 07:34:11.125702 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f9f4c9c8-9kk6b" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 07:34:11 crc kubenswrapper[4960]: I1002 07:34:11.139079 4960 generic.go:334] "Generic (PLEG): container finished" podID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerID="6357085e3742e99be8025a127b2e686d8425f689d1795173cfe1ef5bceb967f2" exitCode=0 Oct 02 07:34:11 crc kubenswrapper[4960]: I1002 07:34:11.139135 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerDied","Data":"6357085e3742e99be8025a127b2e686d8425f689d1795173cfe1ef5bceb967f2"} Oct 02 07:34:12 crc kubenswrapper[4960]: I1002 07:34:12.948717 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.167402 4960 generic.go:334] "Generic (PLEG): container finished" podID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerID="df79319fe38ef1ee5a6208330b070a59bb459cb8673857a3d4c25b3d82c6170d" exitCode=0 Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.167471 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerDied","Data":"df79319fe38ef1ee5a6208330b070a59bb459cb8673857a3d4c25b3d82c6170d"} Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.438433 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76d58bd87b-72mnv" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": dial tcp 10.217.0.156:9311: connect: connection refused" Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.444052 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76d58bd87b-72mnv" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": dial tcp 10.217.0.156:9311: connect: connection refused" Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.445264 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.490213 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.580319 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7854d55f69-7h9wq"] Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.580623 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" podUID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerName="dnsmasq-dns" containerID="cri-o://d9dbb6f4486feaa14174f1bbc5786cd99bc3f06a015aa5db3f9de9c459d6e088" gracePeriod=10 Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.699665 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.161:8776/healthcheck\": dial tcp 10.217.0.161:8776: connect: connection refused" Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.776303 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 07:34:13 crc kubenswrapper[4960]: I1002 07:34:13.860692 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.181798 4960 generic.go:334] "Generic (PLEG): container finished" podID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerID="d9dbb6f4486feaa14174f1bbc5786cd99bc3f06a015aa5db3f9de9c459d6e088" exitCode=0 Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.181896 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" event={"ID":"f8d36716-f1a9-4ce1-8692-65638a5c8884","Type":"ContainerDied","Data":"d9dbb6f4486feaa14174f1bbc5786cd99bc3f06a015aa5db3f9de9c459d6e088"} Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.182182 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="cinder-scheduler" containerID="cri-o://27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649" gracePeriod=30 Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.182308 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="probe" containerID="cri-o://3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc" gracePeriod=30 Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.660779 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9sml2" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.690779 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gqr95" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.774524 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.798159 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5km6\" (UniqueName: \"kubernetes.io/projected/b2a60afc-781b-481f-af68-4d8b9c737d29-kube-api-access-m5km6\") pod \"b2a60afc-781b-481f-af68-4d8b9c737d29\" (UID: \"b2a60afc-781b-481f-af68-4d8b9c737d29\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.798443 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghlwx\" (UniqueName: \"kubernetes.io/projected/54a76d7e-e996-49f6-a3df-7da0d3005b55-kube-api-access-ghlwx\") pod \"54a76d7e-e996-49f6-a3df-7da0d3005b55\" (UID: \"54a76d7e-e996-49f6-a3df-7da0d3005b55\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.805561 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a60afc-781b-481f-af68-4d8b9c737d29-kube-api-access-m5km6" (OuterVolumeSpecName: "kube-api-access-m5km6") pod "b2a60afc-781b-481f-af68-4d8b9c737d29" (UID: "b2a60afc-781b-481f-af68-4d8b9c737d29"). InnerVolumeSpecName "kube-api-access-m5km6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.805783 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a76d7e-e996-49f6-a3df-7da0d3005b55-kube-api-access-ghlwx" (OuterVolumeSpecName: "kube-api-access-ghlwx") pod "54a76d7e-e996-49f6-a3df-7da0d3005b55" (UID: "54a76d7e-e996-49f6-a3df-7da0d3005b55"). InnerVolumeSpecName "kube-api-access-ghlwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900132 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-logs\") pod \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900209 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data-custom\") pod \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900233 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-combined-ca-bundle\") pod \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900336 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-scripts\") pod \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900360 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-etc-machine-id\") pod \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900438 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc6lx\" (UniqueName: \"kubernetes.io/projected/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-kube-api-access-rc6lx\") pod \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900497 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data\") pod \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\" (UID: \"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2\") " Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.900941 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" (UID: "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.901056 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghlwx\" (UniqueName: \"kubernetes.io/projected/54a76d7e-e996-49f6-a3df-7da0d3005b55-kube-api-access-ghlwx\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.901085 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5km6\" (UniqueName: \"kubernetes.io/projected/b2a60afc-781b-481f-af68-4d8b9c737d29-kube-api-access-m5km6\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.901104 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.901465 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-logs" (OuterVolumeSpecName: "logs") pod "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" (UID: "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.907382 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-scripts" (OuterVolumeSpecName: "scripts") pod "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" (UID: "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.914483 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-kube-api-access-rc6lx" (OuterVolumeSpecName: "kube-api-access-rc6lx") pod "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" (UID: "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2"). InnerVolumeSpecName "kube-api-access-rc6lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.920889 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" (UID: "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.971409 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" (UID: "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:14 crc kubenswrapper[4960]: I1002 07:34:14.993300 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.004543 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.004570 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.004582 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.004592 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.004602 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc6lx\" (UniqueName: \"kubernetes.io/projected/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-kube-api-access-rc6lx\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.007103 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data" (OuterVolumeSpecName: "config-data") pod "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" (UID: "ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.105775 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data-custom\") pod \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.106285 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data\") pod \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.106349 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wswm5\" (UniqueName: \"kubernetes.io/projected/8e92f003-5585-452b-9d1e-3915e1ec6f2a-kube-api-access-wswm5\") pod \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.106371 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-combined-ca-bundle\") pod \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.106481 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e92f003-5585-452b-9d1e-3915e1ec6f2a-logs\") pod \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\" (UID: \"8e92f003-5585-452b-9d1e-3915e1ec6f2a\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.106904 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e92f003-5585-452b-9d1e-3915e1ec6f2a-logs" (OuterVolumeSpecName: "logs") pod "8e92f003-5585-452b-9d1e-3915e1ec6f2a" (UID: "8e92f003-5585-452b-9d1e-3915e1ec6f2a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.106997 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.109985 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8e92f003-5585-452b-9d1e-3915e1ec6f2a" (UID: "8e92f003-5585-452b-9d1e-3915e1ec6f2a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.110690 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e92f003-5585-452b-9d1e-3915e1ec6f2a-kube-api-access-wswm5" (OuterVolumeSpecName: "kube-api-access-wswm5") pod "8e92f003-5585-452b-9d1e-3915e1ec6f2a" (UID: "8e92f003-5585-452b-9d1e-3915e1ec6f2a"). InnerVolumeSpecName "kube-api-access-wswm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.120468 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.135042 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.142228 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e92f003-5585-452b-9d1e-3915e1ec6f2a" (UID: "8e92f003-5585-452b-9d1e-3915e1ec6f2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.190269 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data" (OuterVolumeSpecName: "config-data") pod "8e92f003-5585-452b-9d1e-3915e1ec6f2a" (UID: "8e92f003-5585-452b-9d1e-3915e1ec6f2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.196456 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" event={"ID":"f8d36716-f1a9-4ce1-8692-65638a5c8884","Type":"ContainerDied","Data":"8011f7be15fa11b0d8c7d3a81d1953e8bbd1d2f94f0699cdd459b0dcef9763e1"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.196524 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7854d55f69-7h9wq" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.196535 4960 scope.go:117] "RemoveContainer" containerID="d9dbb6f4486feaa14174f1bbc5786cd99bc3f06a015aa5db3f9de9c459d6e088" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.199794 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gqr95" event={"ID":"b2a60afc-781b-481f-af68-4d8b9c737d29","Type":"ContainerDied","Data":"88ed6b96c0ef7df1727212dcd85ab4236e40a97975861b3aacca82d4ebf43663"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.199853 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88ed6b96c0ef7df1727212dcd85ab4236e40a97975861b3aacca82d4ebf43663" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.199936 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gqr95" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207155 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2","Type":"ContainerDied","Data":"5874d4e1a4af7f11ceb99700f3f84b20efdac1f7b52098cb15c8a653b5ff832c"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207209 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207563 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-combined-ca-bundle\") pod \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207671 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-log-httpd\") pod \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207702 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj5kd\" (UniqueName: \"kubernetes.io/projected/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-kube-api-access-mj5kd\") pod \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207774 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-config\") pod \"f8d36716-f1a9-4ce1-8692-65638a5c8884\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207839 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-sb\") pod \"f8d36716-f1a9-4ce1-8692-65638a5c8884\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207872 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-config-data\") pod \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207911 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffp6c\" (UniqueName: \"kubernetes.io/projected/f8d36716-f1a9-4ce1-8692-65638a5c8884-kube-api-access-ffp6c\") pod \"f8d36716-f1a9-4ce1-8692-65638a5c8884\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.207954 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-nb\") pod \"f8d36716-f1a9-4ce1-8692-65638a5c8884\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208052 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-dns-svc\") pod \"f8d36716-f1a9-4ce1-8692-65638a5c8884\" (UID: \"f8d36716-f1a9-4ce1-8692-65638a5c8884\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208099 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-sg-core-conf-yaml\") pod \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208124 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-run-httpd\") pod \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208240 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-scripts\") pod \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\" (UID: \"60cb54d3-3ec5-4be0-a050-6dd637db1c2e\") " Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208418 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "60cb54d3-3ec5-4be0-a050-6dd637db1c2e" (UID: "60cb54d3-3ec5-4be0-a050-6dd637db1c2e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208666 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e92f003-5585-452b-9d1e-3915e1ec6f2a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208680 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208694 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208705 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208716 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wswm5\" (UniqueName: \"kubernetes.io/projected/8e92f003-5585-452b-9d1e-3915e1ec6f2a-kube-api-access-wswm5\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.208725 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e92f003-5585-452b-9d1e-3915e1ec6f2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.213588 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-kube-api-access-mj5kd" (OuterVolumeSpecName: "kube-api-access-mj5kd") pod "60cb54d3-3ec5-4be0-a050-6dd637db1c2e" (UID: "60cb54d3-3ec5-4be0-a050-6dd637db1c2e"). InnerVolumeSpecName "kube-api-access-mj5kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.213861 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "60cb54d3-3ec5-4be0-a050-6dd637db1c2e" (UID: "60cb54d3-3ec5-4be0-a050-6dd637db1c2e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.216874 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60cb54d3-3ec5-4be0-a050-6dd637db1c2e","Type":"ContainerDied","Data":"42dbc9a9b9bac9d8dd7aa5b717f3c17cf173c8b9a173820dc5fdc375dd38e4e1"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.216908 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.221305 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9sml2" event={"ID":"54a76d7e-e996-49f6-a3df-7da0d3005b55","Type":"ContainerDied","Data":"9149b5d70188c5dbcd23e19fa93774f64f9ad7ff370776469ce7d452cb48eec1"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.221334 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9149b5d70188c5dbcd23e19fa93774f64f9ad7ff370776469ce7d452cb48eec1" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.221346 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9sml2" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.222292 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-scripts" (OuterVolumeSpecName: "scripts") pod "60cb54d3-3ec5-4be0-a050-6dd637db1c2e" (UID: "60cb54d3-3ec5-4be0-a050-6dd637db1c2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.222380 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d36716-f1a9-4ce1-8692-65638a5c8884-kube-api-access-ffp6c" (OuterVolumeSpecName: "kube-api-access-ffp6c") pod "f8d36716-f1a9-4ce1-8692-65638a5c8884" (UID: "f8d36716-f1a9-4ce1-8692-65638a5c8884"). InnerVolumeSpecName "kube-api-access-ffp6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.222739 4960 scope.go:117] "RemoveContainer" containerID="7409f5d4cb1f48fb62b1a5358be6470cc3b12ddfa36d61bd06a75f778e5b9b69" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.223009 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d58bd87b-72mnv" event={"ID":"8e92f003-5585-452b-9d1e-3915e1ec6f2a","Type":"ContainerDied","Data":"7d09bfbbf318e6776baddda85a65dd748258ab7de9a51650681ca2bb171dff55"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.223061 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76d58bd87b-72mnv" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.226404 4960 generic.go:334] "Generic (PLEG): container finished" podID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerID="3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc" exitCode=0 Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.226472 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31","Type":"ContainerDied","Data":"3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.228630 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"34704ecb-b08d-4f67-b6ed-66141a0f6e74","Type":"ContainerStarted","Data":"46a4b15692f60503d5cbebbe6f67cd26593830c619b434653524ccc1f2ae835d"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.229791 4960 generic.go:334] "Generic (PLEG): container finished" podID="61aefc93-ba77-48df-af65-af010d8a8efa" containerID="3b1a2da3078786d586b932d6f731e9eba3090a4df06071a1bc98faa5f1ceff63" exitCode=0 Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.229828 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tqvz5" event={"ID":"61aefc93-ba77-48df-af65-af010d8a8efa","Type":"ContainerDied","Data":"3b1a2da3078786d586b932d6f731e9eba3090a4df06071a1bc98faa5f1ceff63"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.229846 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tqvz5" event={"ID":"61aefc93-ba77-48df-af65-af010d8a8efa","Type":"ContainerStarted","Data":"51884c8525ac7c4e523ef36023be1269af5ffcf57f97dcdf7f0fcf4d18f1bb66"} Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.266578 4960 scope.go:117] "RemoveContainer" containerID="17710c9519ac7c5849f4e05184b02837ebbaad2552ae1cb05f8014b63c3086e1" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.284079 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "60cb54d3-3ec5-4be0-a050-6dd637db1c2e" (UID: "60cb54d3-3ec5-4be0-a050-6dd637db1c2e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.293914 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.299839 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f8d36716-f1a9-4ce1-8692-65638a5c8884" (UID: "f8d36716-f1a9-4ce1-8692-65638a5c8884"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.305894 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.308221 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f8d36716-f1a9-4ce1-8692-65638a5c8884" (UID: "f8d36716-f1a9-4ce1-8692-65638a5c8884"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.310958 4960 scope.go:117] "RemoveContainer" containerID="247beb3ea514f9ee887529c686fcab5cf90a7fe5061f815efe48fac3e133b9a3" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.318507 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-config" (OuterVolumeSpecName: "config") pod "f8d36716-f1a9-4ce1-8692-65638a5c8884" (UID: "f8d36716-f1a9-4ce1-8692-65638a5c8884"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319503 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319529 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj5kd\" (UniqueName: \"kubernetes.io/projected/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-kube-api-access-mj5kd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319545 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319557 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319570 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffp6c\" (UniqueName: \"kubernetes.io/projected/f8d36716-f1a9-4ce1-8692-65638a5c8884-kube-api-access-ffp6c\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319583 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319598 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.319609 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.326618 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327659 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-central-agent" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327677 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-central-agent" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327697 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api-log" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327705 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api-log" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327713 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="sg-core" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327720 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="sg-core" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327729 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-notification-agent" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327735 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-notification-agent" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327762 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a60afc-781b-481f-af68-4d8b9c737d29" containerName="mariadb-database-create" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327768 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a60afc-781b-481f-af68-4d8b9c737d29" containerName="mariadb-database-create" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327779 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="proxy-httpd" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327787 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="proxy-httpd" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327801 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327810 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327822 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327828 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327839 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a76d7e-e996-49f6-a3df-7da0d3005b55" containerName="mariadb-database-create" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327845 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a76d7e-e996-49f6-a3df-7da0d3005b55" containerName="mariadb-database-create" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327857 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api-log" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327863 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api-log" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327874 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerName="init" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327880 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerName="init" Oct 02 07:34:15 crc kubenswrapper[4960]: E1002 07:34:15.327889 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerName="dnsmasq-dns" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.327895 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerName="dnsmasq-dns" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328112 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d36716-f1a9-4ce1-8692-65638a5c8884" containerName="dnsmasq-dns" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328125 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api-log" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328141 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-notification-agent" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328162 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api-log" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328173 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a60afc-781b-481f-af68-4d8b9c737d29" containerName="mariadb-database-create" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328184 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="ceilometer-central-agent" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328195 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a76d7e-e996-49f6-a3df-7da0d3005b55" containerName="mariadb-database-create" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328209 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" containerName="barbican-api" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328219 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" containerName="cinder-api" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328228 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="sg-core" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.328242 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" containerName="proxy-httpd" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.331726 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.334460 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.334558 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.334613 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.351777 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.361687 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f8d36716-f1a9-4ce1-8692-65638a5c8884" (UID: "f8d36716-f1a9-4ce1-8692-65638a5c8884"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.364024 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.670119389 podStartE2EDuration="20.363996406s" podCreationTimestamp="2025-10-02 07:33:55 +0000 UTC" firstStartedPulling="2025-10-02 07:33:56.916072338 +0000 UTC m=+1057.948018635" lastFinishedPulling="2025-10-02 07:34:14.609949365 +0000 UTC m=+1075.641895652" observedRunningTime="2025-10-02 07:34:15.310581247 +0000 UTC m=+1076.342527534" watchObservedRunningTime="2025-10-02 07:34:15.363996406 +0000 UTC m=+1076.395942693" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.377116 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-76d58bd87b-72mnv"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.377927 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d8fb8597c-r7rqs" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.383648 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-76d58bd87b-72mnv"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.385086 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-config-data" (OuterVolumeSpecName: "config-data") pod "60cb54d3-3ec5-4be0-a050-6dd637db1c2e" (UID: "60cb54d3-3ec5-4be0-a050-6dd637db1c2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.408107 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60cb54d3-3ec5-4be0-a050-6dd637db1c2e" (UID: "60cb54d3-3ec5-4be0-a050-6dd637db1c2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.422998 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnrkn\" (UniqueName: \"kubernetes.io/projected/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-kube-api-access-qnrkn\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.423051 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-scripts\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.423389 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-config-data-custom\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.423558 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-public-tls-certs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.423677 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.423800 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-logs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.423919 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-config-data\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.423953 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-etc-machine-id\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.424053 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.424129 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8d36716-f1a9-4ce1-8692-65638a5c8884-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.424143 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.424156 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cb54d3-3ec5-4be0-a050-6dd637db1c2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.450023 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ff76df558-2w955"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.451515 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6ff76df558-2w955" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-api" containerID="cri-o://665ef4457fcb60e709404f75b3aa015ad8af4cd231ac3754feb33527c9f3946e" gracePeriod=30 Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.452037 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6ff76df558-2w955" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-httpd" containerID="cri-o://a33ccfa61b073f58fe5079e93ef170a2a45cb9a82220d4f291da1ff7065f246f" gracePeriod=30 Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.527813 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-config-data\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.527869 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-etc-machine-id\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.527910 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.527948 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnrkn\" (UniqueName: \"kubernetes.io/projected/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-kube-api-access-qnrkn\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.527991 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-scripts\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.528113 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-config-data-custom\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.528155 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-public-tls-certs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.528208 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.528295 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-logs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.529543 4960 scope.go:117] "RemoveContainer" containerID="9e96d525fba70d1ccbdc935eea63997f47269c6b8a4a58beb2523f88d1d262bd" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.530090 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-logs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.530131 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-etc-machine-id\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.536578 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-config-data\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.538474 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.541634 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-scripts\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.544301 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.545534 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7854d55f69-7h9wq"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.548321 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-public-tls-certs\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.550107 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnrkn\" (UniqueName: \"kubernetes.io/projected/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-kube-api-access-qnrkn\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.571910 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7854d55f69-7h9wq"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.572495 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/913ab46c-3d2d-407c-bd43-6f4b9f9e8f58-config-data-custom\") pod \"cinder-api-0\" (UID: \"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58\") " pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.637612 4960 scope.go:117] "RemoveContainer" containerID="199fc2d4ad72ce0da2819fc6e6feaddc57685dc00eae6739440498a42c69e902" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.655327 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.665515 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.670940 4960 scope.go:117] "RemoveContainer" containerID="df79319fe38ef1ee5a6208330b070a59bb459cb8673857a3d4c25b3d82c6170d" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.678034 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.681158 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.703568 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.703934 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.713956 4960 scope.go:117] "RemoveContainer" containerID="6357085e3742e99be8025a127b2e686d8425f689d1795173cfe1ef5bceb967f2" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.715264 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.742572 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phlt8\" (UniqueName: \"kubernetes.io/projected/efea504b-eb3a-41e9-804c-021e736dfd61-kube-api-access-phlt8\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.742655 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.742703 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-config-data\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.742787 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-run-httpd\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.742812 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.742834 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-scripts\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.742855 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-log-httpd\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.778236 4960 scope.go:117] "RemoveContainer" containerID="49b602fa88e894c950210ca54cd12180f1f11b32cfd783e22adda8ef2b9bf414" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.807498 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.826107 4960 scope.go:117] "RemoveContainer" containerID="681877e74714d583554fe8d657317e977c5760cde38447998da93ccb223a0623" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.845552 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-config-data\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.846425 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-run-httpd\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.847512 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-run-httpd\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.847713 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.847796 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-scripts\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.847905 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-log-httpd\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.848048 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phlt8\" (UniqueName: \"kubernetes.io/projected/efea504b-eb3a-41e9-804c-021e736dfd61-kube-api-access-phlt8\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.848146 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.851008 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-log-httpd\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.854317 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.857511 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.857845 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-config-data\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.862546 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-scripts\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:15 crc kubenswrapper[4960]: I1002 07:34:15.868147 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phlt8\" (UniqueName: \"kubernetes.io/projected/efea504b-eb3a-41e9-804c-021e736dfd61-kube-api-access-phlt8\") pod \"ceilometer-0\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " pod="openstack/ceilometer-0" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.061942 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.255928 4960 generic.go:334] "Generic (PLEG): container finished" podID="12495902-a857-4ea0-b5bc-178803f33c12" containerID="18bae37076d32c0f6e54e8ed4eb2151d38367a0781878ae2c2c5ffe57acd4e5f" exitCode=137 Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.256024 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f9f4c9c8-9kk6b" event={"ID":"12495902-a857-4ea0-b5bc-178803f33c12","Type":"ContainerDied","Data":"18bae37076d32c0f6e54e8ed4eb2151d38367a0781878ae2c2c5ffe57acd4e5f"} Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.261732 4960 generic.go:334] "Generic (PLEG): container finished" podID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerID="a33ccfa61b073f58fe5079e93ef170a2a45cb9a82220d4f291da1ff7065f246f" exitCode=0 Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.262923 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ff76df558-2w955" event={"ID":"19b6c4e4-068d-4969-a8b3-95d907bde1c5","Type":"ContainerDied","Data":"a33ccfa61b073f58fe5079e93ef170a2a45cb9a82220d4f291da1ff7065f246f"} Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.355716 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60cb54d3-3ec5-4be0-a050-6dd637db1c2e" path="/var/lib/kubelet/pods/60cb54d3-3ec5-4be0-a050-6dd637db1c2e/volumes" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.356656 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e92f003-5585-452b-9d1e-3915e1ec6f2a" path="/var/lib/kubelet/pods/8e92f003-5585-452b-9d1e-3915e1ec6f2a/volumes" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.358100 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2" path="/var/lib/kubelet/pods/ac1f219d-59a3-4c0c-ab27-03b8b3fca7e2/volumes" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.359296 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d36716-f1a9-4ce1-8692-65638a5c8884" path="/var/lib/kubelet/pods/f8d36716-f1a9-4ce1-8692-65638a5c8884/volumes" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.360082 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.561885 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.668276 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:16 crc kubenswrapper[4960]: W1002 07:34:16.677581 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefea504b_eb3a_41e9_804c_021e736dfd61.slice/crio-21e4621e38f7545737d3b9386eea83dea7b18473c1ee9527038f859328f2a513 WatchSource:0}: Error finding container 21e4621e38f7545737d3b9386eea83dea7b18473c1ee9527038f859328f2a513: Status 404 returned error can't find the container with id 21e4621e38f7545737d3b9386eea83dea7b18473c1ee9527038f859328f2a513 Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.678798 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-tls-certs\") pod \"12495902-a857-4ea0-b5bc-178803f33c12\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.678917 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-secret-key\") pod \"12495902-a857-4ea0-b5bc-178803f33c12\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.679028 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12495902-a857-4ea0-b5bc-178803f33c12-logs\") pod \"12495902-a857-4ea0-b5bc-178803f33c12\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.679111 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-scripts\") pod \"12495902-a857-4ea0-b5bc-178803f33c12\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.679181 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-combined-ca-bundle\") pod \"12495902-a857-4ea0-b5bc-178803f33c12\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.679281 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-config-data\") pod \"12495902-a857-4ea0-b5bc-178803f33c12\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.679313 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvsmj\" (UniqueName: \"kubernetes.io/projected/12495902-a857-4ea0-b5bc-178803f33c12-kube-api-access-qvsmj\") pod \"12495902-a857-4ea0-b5bc-178803f33c12\" (UID: \"12495902-a857-4ea0-b5bc-178803f33c12\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.681142 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12495902-a857-4ea0-b5bc-178803f33c12-logs" (OuterVolumeSpecName: "logs") pod "12495902-a857-4ea0-b5bc-178803f33c12" (UID: "12495902-a857-4ea0-b5bc-178803f33c12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.691178 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "12495902-a857-4ea0-b5bc-178803f33c12" (UID: "12495902-a857-4ea0-b5bc-178803f33c12"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.691204 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12495902-a857-4ea0-b5bc-178803f33c12-kube-api-access-qvsmj" (OuterVolumeSpecName: "kube-api-access-qvsmj") pod "12495902-a857-4ea0-b5bc-178803f33c12" (UID: "12495902-a857-4ea0-b5bc-178803f33c12"). InnerVolumeSpecName "kube-api-access-qvsmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.707357 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqvz5" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.712767 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-config-data" (OuterVolumeSpecName: "config-data") pod "12495902-a857-4ea0-b5bc-178803f33c12" (UID: "12495902-a857-4ea0-b5bc-178803f33c12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.731250 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12495902-a857-4ea0-b5bc-178803f33c12" (UID: "12495902-a857-4ea0-b5bc-178803f33c12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.739654 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-scripts" (OuterVolumeSpecName: "scripts") pod "12495902-a857-4ea0-b5bc-178803f33c12" (UID: "12495902-a857-4ea0-b5bc-178803f33c12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.761537 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "12495902-a857-4ea0-b5bc-178803f33c12" (UID: "12495902-a857-4ea0-b5bc-178803f33c12"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783336 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwgbh\" (UniqueName: \"kubernetes.io/projected/61aefc93-ba77-48df-af65-af010d8a8efa-kube-api-access-fwgbh\") pod \"61aefc93-ba77-48df-af65-af010d8a8efa\" (UID: \"61aefc93-ba77-48df-af65-af010d8a8efa\") " Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783737 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783753 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvsmj\" (UniqueName: \"kubernetes.io/projected/12495902-a857-4ea0-b5bc-178803f33c12-kube-api-access-qvsmj\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783762 4960 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783772 4960 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783780 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12495902-a857-4ea0-b5bc-178803f33c12-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783788 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12495902-a857-4ea0-b5bc-178803f33c12-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.783796 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12495902-a857-4ea0-b5bc-178803f33c12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.787172 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61aefc93-ba77-48df-af65-af010d8a8efa-kube-api-access-fwgbh" (OuterVolumeSpecName: "kube-api-access-fwgbh") pod "61aefc93-ba77-48df-af65-af010d8a8efa" (UID: "61aefc93-ba77-48df-af65-af010d8a8efa"). InnerVolumeSpecName "kube-api-access-fwgbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:16 crc kubenswrapper[4960]: I1002 07:34:16.889993 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwgbh\" (UniqueName: \"kubernetes.io/projected/61aefc93-ba77-48df-af65-af010d8a8efa-kube-api-access-fwgbh\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.276602 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tqvz5" event={"ID":"61aefc93-ba77-48df-af65-af010d8a8efa","Type":"ContainerDied","Data":"51884c8525ac7c4e523ef36023be1269af5ffcf57f97dcdf7f0fcf4d18f1bb66"} Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.277035 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51884c8525ac7c4e523ef36023be1269af5ffcf57f97dcdf7f0fcf4d18f1bb66" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.276757 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqvz5" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.284229 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerStarted","Data":"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554"} Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.284296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerStarted","Data":"21e4621e38f7545737d3b9386eea83dea7b18473c1ee9527038f859328f2a513"} Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.286396 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58","Type":"ContainerStarted","Data":"63192c43b53a729c1a01e9081a75d684ffc478e25316506c63a05e018ae5b1d8"} Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.286438 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58","Type":"ContainerStarted","Data":"4a61e2eb49d33d01bdd3dec1a3bcb506f180d4483da7a5df88e54a0906433e5c"} Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.293661 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f9f4c9c8-9kk6b" event={"ID":"12495902-a857-4ea0-b5bc-178803f33c12","Type":"ContainerDied","Data":"84a535d2f7b6fb0a202941ae86960bc22df01843f5041f77fb8b72f85460edc8"} Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.293712 4960 scope.go:117] "RemoveContainer" containerID="577ac4501e9a0ce39b630127d7651cd4d3c5b26e15121bfe0333a8b403d603b5" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.294004 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f9f4c9c8-9kk6b" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.302710 4960 generic.go:334] "Generic (PLEG): container finished" podID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerID="665ef4457fcb60e709404f75b3aa015ad8af4cd231ac3754feb33527c9f3946e" exitCode=0 Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.302786 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ff76df558-2w955" event={"ID":"19b6c4e4-068d-4969-a8b3-95d907bde1c5","Type":"ContainerDied","Data":"665ef4457fcb60e709404f75b3aa015ad8af4cd231ac3754feb33527c9f3946e"} Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.371334 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85f9f4c9c8-9kk6b"] Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.378428 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85f9f4c9c8-9kk6b"] Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.533265 4960 scope.go:117] "RemoveContainer" containerID="18bae37076d32c0f6e54e8ed4eb2151d38367a0781878ae2c2c5ffe57acd4e5f" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.756040 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.809373 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-ovndb-tls-certs\") pod \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.809524 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-config\") pod \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.809576 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmrw8\" (UniqueName: \"kubernetes.io/projected/19b6c4e4-068d-4969-a8b3-95d907bde1c5-kube-api-access-zmrw8\") pod \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.809663 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-combined-ca-bundle\") pod \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.809853 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-httpd-config\") pod \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\" (UID: \"19b6c4e4-068d-4969-a8b3-95d907bde1c5\") " Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.817390 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "19b6c4e4-068d-4969-a8b3-95d907bde1c5" (UID: "19b6c4e4-068d-4969-a8b3-95d907bde1c5"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.817469 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b6c4e4-068d-4969-a8b3-95d907bde1c5-kube-api-access-zmrw8" (OuterVolumeSpecName: "kube-api-access-zmrw8") pod "19b6c4e4-068d-4969-a8b3-95d907bde1c5" (UID: "19b6c4e4-068d-4969-a8b3-95d907bde1c5"). InnerVolumeSpecName "kube-api-access-zmrw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.912924 4960 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.913438 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmrw8\" (UniqueName: \"kubernetes.io/projected/19b6c4e4-068d-4969-a8b3-95d907bde1c5-kube-api-access-zmrw8\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.927964 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "19b6c4e4-068d-4969-a8b3-95d907bde1c5" (UID: "19b6c4e4-068d-4969-a8b3-95d907bde1c5"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.936288 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-config" (OuterVolumeSpecName: "config") pod "19b6c4e4-068d-4969-a8b3-95d907bde1c5" (UID: "19b6c4e4-068d-4969-a8b3-95d907bde1c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:17 crc kubenswrapper[4960]: I1002 07:34:17.970425 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19b6c4e4-068d-4969-a8b3-95d907bde1c5" (UID: "19b6c4e4-068d-4969-a8b3-95d907bde1c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.015204 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.015486 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.015502 4960 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b6c4e4-068d-4969-a8b3-95d907bde1c5-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.319611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ff76df558-2w955" event={"ID":"19b6c4e4-068d-4969-a8b3-95d907bde1c5","Type":"ContainerDied","Data":"57042676f06a48c73dcb789fde695ef5f7d352e31e62f782c3b084c640dcc8ce"} Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.320195 4960 scope.go:117] "RemoveContainer" containerID="a33ccfa61b073f58fe5079e93ef170a2a45cb9a82220d4f291da1ff7065f246f" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.320253 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.355451 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12495902-a857-4ea0-b5bc-178803f33c12" path="/var/lib/kubelet/pods/12495902-a857-4ea0-b5bc-178803f33c12/volumes" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.356967 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.357016 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerStarted","Data":"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a"} Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.357034 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"913ab46c-3d2d-407c-bd43-6f4b9f9e8f58","Type":"ContainerStarted","Data":"ae0687d8c0f8a6756c34e17ce70b5f932ce1d5b631e043b6f21f75ca821b2734"} Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.370864 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.370839877 podStartE2EDuration="3.370839877s" podCreationTimestamp="2025-10-02 07:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:18.359046593 +0000 UTC m=+1079.390992880" watchObservedRunningTime="2025-10-02 07:34:18.370839877 +0000 UTC m=+1079.402786174" Oct 02 07:34:18 crc kubenswrapper[4960]: I1002 07:34:18.525170 4960 scope.go:117] "RemoveContainer" containerID="665ef4457fcb60e709404f75b3aa015ad8af4cd231ac3754feb33527c9f3946e" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.242910 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.341759 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-scripts\") pod \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.341853 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data\") pod \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.342135 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data-custom\") pod \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.342169 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-combined-ca-bundle\") pod \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.342201 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-etc-machine-id\") pod \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.342277 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2kvp\" (UniqueName: \"kubernetes.io/projected/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-kube-api-access-j2kvp\") pod \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\" (UID: \"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31\") " Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.350614 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" (UID: "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.364575 4960 generic.go:334] "Generic (PLEG): container finished" podID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerID="27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649" exitCode=0 Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.364697 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.364720 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" (UID: "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.364805 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31","Type":"ContainerDied","Data":"27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649"} Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.364850 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5e02dcc-6bf6-4e1d-b560-234f6ff46d31","Type":"ContainerDied","Data":"82e7f6b11a4a91d0b4a10b333b14882778b78308dbb5110752da74f7d2fed0fe"} Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.364898 4960 scope.go:117] "RemoveContainer" containerID="3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.369717 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-scripts" (OuterVolumeSpecName: "scripts") pod "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" (UID: "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.374287 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-kube-api-access-j2kvp" (OuterVolumeSpecName: "kube-api-access-j2kvp") pod "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" (UID: "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31"). InnerVolumeSpecName "kube-api-access-j2kvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.374419 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerStarted","Data":"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db"} Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.427252 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" (UID: "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.444642 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.444679 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.444691 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.444701 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2kvp\" (UniqueName: \"kubernetes.io/projected/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-kube-api-access-j2kvp\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.444711 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.506058 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data" (OuterVolumeSpecName: "config-data") pod "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" (UID: "c5e02dcc-6bf6-4e1d-b560-234f6ff46d31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.552536 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.615239 4960 scope.go:117] "RemoveContainer" containerID="27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.636824 4960 scope.go:117] "RemoveContainer" containerID="3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.640761 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc\": container with ID starting with 3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc not found: ID does not exist" containerID="3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.640822 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc"} err="failed to get container status \"3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc\": rpc error: code = NotFound desc = could not find container \"3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc\": container with ID starting with 3e08334312a8fd5c4814ee00bc87356d8daf53a55d47781878a048ad18dab4dc not found: ID does not exist" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.640859 4960 scope.go:117] "RemoveContainer" containerID="27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.641538 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649\": container with ID starting with 27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649 not found: ID does not exist" containerID="27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.641597 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649"} err="failed to get container status \"27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649\": rpc error: code = NotFound desc = could not find container \"27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649\": container with ID starting with 27fd0015bcb6cd6cda30093b10b4fc39d2479bbfba89681b9726ce77446e9649 not found: ID does not exist" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.701278 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.710325 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.739869 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.740470 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="probe" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740499 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="probe" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.740531 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61aefc93-ba77-48df-af65-af010d8a8efa" containerName="mariadb-database-create" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740540 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="61aefc93-ba77-48df-af65-af010d8a8efa" containerName="mariadb-database-create" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.740552 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-httpd" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740565 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-httpd" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.740580 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="cinder-scheduler" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740589 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="cinder-scheduler" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.740601 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740608 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.740630 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon-log" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740638 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon-log" Oct 02 07:34:19 crc kubenswrapper[4960]: E1002 07:34:19.740664 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-api" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740675 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-api" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740893 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740909 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="probe" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740929 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="61aefc93-ba77-48df-af65-af010d8a8efa" containerName="mariadb-database-create" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740962 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="12495902-a857-4ea0-b5bc-178803f33c12" containerName="horizon-log" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740991 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" containerName="cinder-scheduler" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.740999 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-httpd" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.741010 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" containerName="neutron-api" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.742381 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.746329 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.756191 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.859021 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-scripts\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.859428 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t9v7\" (UniqueName: \"kubernetes.io/projected/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-kube-api-access-5t9v7\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.859562 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.859709 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-config-data\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.859840 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.860015 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.961942 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-scripts\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.962018 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t9v7\" (UniqueName: \"kubernetes.io/projected/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-kube-api-access-5t9v7\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.962051 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.962095 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-config-data\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.962179 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.962213 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.962500 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.966750 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-scripts\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.966813 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.967034 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-config-data\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.967468 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:19 crc kubenswrapper[4960]: I1002 07:34:19.979671 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t9v7\" (UniqueName: \"kubernetes.io/projected/f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d-kube-api-access-5t9v7\") pod \"cinder-scheduler-0\" (UID: \"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d\") " pod="openstack/cinder-scheduler-0" Oct 02 07:34:20 crc kubenswrapper[4960]: I1002 07:34:20.072121 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:34:20 crc kubenswrapper[4960]: I1002 07:34:20.343303 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5e02dcc-6bf6-4e1d-b560-234f6ff46d31" path="/var/lib/kubelet/pods/c5e02dcc-6bf6-4e1d-b560-234f6ff46d31/volumes" Oct 02 07:34:20 crc kubenswrapper[4960]: I1002 07:34:20.418094 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerStarted","Data":"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3"} Oct 02 07:34:20 crc kubenswrapper[4960]: I1002 07:34:20.418364 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:34:20 crc kubenswrapper[4960]: I1002 07:34:20.535002 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.676370516 podStartE2EDuration="5.534959208s" podCreationTimestamp="2025-10-02 07:34:15 +0000 UTC" firstStartedPulling="2025-10-02 07:34:16.693171093 +0000 UTC m=+1077.725117380" lastFinishedPulling="2025-10-02 07:34:19.551759785 +0000 UTC m=+1080.583706072" observedRunningTime="2025-10-02 07:34:20.450563943 +0000 UTC m=+1081.482510240" watchObservedRunningTime="2025-10-02 07:34:20.534959208 +0000 UTC m=+1081.566905495" Oct 02 07:34:20 crc kubenswrapper[4960]: I1002 07:34:20.540534 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:34:20 crc kubenswrapper[4960]: W1002 07:34:20.543650 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4c3ac70_8228_472d_90e9_dc3ab6fb6e5d.slice/crio-695fd521748913649b711b17c7daa4e09bcbab1817042c2dd39339d2565147bf WatchSource:0}: Error finding container 695fd521748913649b711b17c7daa4e09bcbab1817042c2dd39339d2565147bf: Status 404 returned error can't find the container with id 695fd521748913649b711b17c7daa4e09bcbab1817042c2dd39339d2565147bf Oct 02 07:34:21 crc kubenswrapper[4960]: I1002 07:34:21.441657 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d","Type":"ContainerStarted","Data":"3e5a494f55ebbc83733913155ea933a035e2cdc3f90bab80e639af1b3aa14ba2"} Oct 02 07:34:21 crc kubenswrapper[4960]: I1002 07:34:21.442564 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d","Type":"ContainerStarted","Data":"695fd521748913649b711b17c7daa4e09bcbab1817042c2dd39339d2565147bf"} Oct 02 07:34:22 crc kubenswrapper[4960]: I1002 07:34:22.452010 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d","Type":"ContainerStarted","Data":"cfc67561f22659555fee797142f0821b460a63cf815dac863f9617fba9093501"} Oct 02 07:34:22 crc kubenswrapper[4960]: I1002 07:34:22.483305 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.483282426 podStartE2EDuration="3.483282426s" podCreationTimestamp="2025-10-02 07:34:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:22.478693682 +0000 UTC m=+1083.510639969" watchObservedRunningTime="2025-10-02 07:34:22.483282426 +0000 UTC m=+1083.515228713" Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.095169 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.095490 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-central-agent" containerID="cri-o://022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554" gracePeriod=30 Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.095636 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="proxy-httpd" containerID="cri-o://bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3" gracePeriod=30 Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.095684 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="sg-core" containerID="cri-o://40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db" gracePeriod=30 Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.095809 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-notification-agent" containerID="cri-o://172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a" gracePeriod=30 Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.462508 4960 generic.go:334] "Generic (PLEG): container finished" podID="efea504b-eb3a-41e9-804c-021e736dfd61" containerID="bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3" exitCode=0 Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.462906 4960 generic.go:334] "Generic (PLEG): container finished" podID="efea504b-eb3a-41e9-804c-021e736dfd61" containerID="40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db" exitCode=2 Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.462589 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerDied","Data":"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3"} Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.462990 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerDied","Data":"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db"} Oct 02 07:34:23 crc kubenswrapper[4960]: I1002 07:34:23.981226 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.052959 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-config-data\") pod \"efea504b-eb3a-41e9-804c-021e736dfd61\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.053051 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-sg-core-conf-yaml\") pod \"efea504b-eb3a-41e9-804c-021e736dfd61\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.053133 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phlt8\" (UniqueName: \"kubernetes.io/projected/efea504b-eb3a-41e9-804c-021e736dfd61-kube-api-access-phlt8\") pod \"efea504b-eb3a-41e9-804c-021e736dfd61\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.053213 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-run-httpd\") pod \"efea504b-eb3a-41e9-804c-021e736dfd61\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.053290 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-combined-ca-bundle\") pod \"efea504b-eb3a-41e9-804c-021e736dfd61\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.053402 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-scripts\") pod \"efea504b-eb3a-41e9-804c-021e736dfd61\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.053546 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-log-httpd\") pod \"efea504b-eb3a-41e9-804c-021e736dfd61\" (UID: \"efea504b-eb3a-41e9-804c-021e736dfd61\") " Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.054624 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "efea504b-eb3a-41e9-804c-021e736dfd61" (UID: "efea504b-eb3a-41e9-804c-021e736dfd61"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.056309 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "efea504b-eb3a-41e9-804c-021e736dfd61" (UID: "efea504b-eb3a-41e9-804c-021e736dfd61"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.061749 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-scripts" (OuterVolumeSpecName: "scripts") pod "efea504b-eb3a-41e9-804c-021e736dfd61" (UID: "efea504b-eb3a-41e9-804c-021e736dfd61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.061971 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efea504b-eb3a-41e9-804c-021e736dfd61-kube-api-access-phlt8" (OuterVolumeSpecName: "kube-api-access-phlt8") pod "efea504b-eb3a-41e9-804c-021e736dfd61" (UID: "efea504b-eb3a-41e9-804c-021e736dfd61"). InnerVolumeSpecName "kube-api-access-phlt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.104377 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "efea504b-eb3a-41e9-804c-021e736dfd61" (UID: "efea504b-eb3a-41e9-804c-021e736dfd61"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.156075 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.156113 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.156123 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.156135 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phlt8\" (UniqueName: \"kubernetes.io/projected/efea504b-eb3a-41e9-804c-021e736dfd61-kube-api-access-phlt8\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.156146 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efea504b-eb3a-41e9-804c-021e736dfd61-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.162010 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-config-data" (OuterVolumeSpecName: "config-data") pod "efea504b-eb3a-41e9-804c-021e736dfd61" (UID: "efea504b-eb3a-41e9-804c-021e736dfd61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.171450 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efea504b-eb3a-41e9-804c-021e736dfd61" (UID: "efea504b-eb3a-41e9-804c-021e736dfd61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.262870 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.262920 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efea504b-eb3a-41e9-804c-021e736dfd61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.475498 4960 generic.go:334] "Generic (PLEG): container finished" podID="efea504b-eb3a-41e9-804c-021e736dfd61" containerID="172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a" exitCode=0 Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.475538 4960 generic.go:334] "Generic (PLEG): container finished" podID="efea504b-eb3a-41e9-804c-021e736dfd61" containerID="022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554" exitCode=0 Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.475562 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerDied","Data":"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a"} Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.475599 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerDied","Data":"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554"} Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.475611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efea504b-eb3a-41e9-804c-021e736dfd61","Type":"ContainerDied","Data":"21e4621e38f7545737d3b9386eea83dea7b18473c1ee9527038f859328f2a513"} Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.475630 4960 scope.go:117] "RemoveContainer" containerID="bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.475681 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.516388 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.518302 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.536339 4960 scope.go:117] "RemoveContainer" containerID="40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.537651 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.538105 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-notification-agent" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538124 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-notification-agent" Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.538143 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-central-agent" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538150 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-central-agent" Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.538159 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="sg-core" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538167 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="sg-core" Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.538177 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="proxy-httpd" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538183 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="proxy-httpd" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538336 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-central-agent" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538354 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="proxy-httpd" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538371 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="sg-core" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.538383 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" containerName="ceilometer-notification-agent" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.539991 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.544478 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.545113 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.563730 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.568520 4960 scope.go:117] "RemoveContainer" containerID="172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.571002 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.571072 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-scripts\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.571184 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-log-httpd\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.571282 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.571455 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztxg7\" (UniqueName: \"kubernetes.io/projected/433e660d-96fa-4146-998d-dd2f36a64e32-kube-api-access-ztxg7\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.571516 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-run-httpd\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.571617 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-config-data\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.603029 4960 scope.go:117] "RemoveContainer" containerID="022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.629768 4960 scope.go:117] "RemoveContainer" containerID="bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3" Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.630464 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3\": container with ID starting with bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3 not found: ID does not exist" containerID="bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.630499 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3"} err="failed to get container status \"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3\": rpc error: code = NotFound desc = could not find container \"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3\": container with ID starting with bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3 not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.630525 4960 scope.go:117] "RemoveContainer" containerID="40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db" Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.630833 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db\": container with ID starting with 40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db not found: ID does not exist" containerID="40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.630872 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db"} err="failed to get container status \"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db\": rpc error: code = NotFound desc = could not find container \"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db\": container with ID starting with 40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.630892 4960 scope.go:117] "RemoveContainer" containerID="172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a" Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.631161 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a\": container with ID starting with 172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a not found: ID does not exist" containerID="172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.631202 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a"} err="failed to get container status \"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a\": rpc error: code = NotFound desc = could not find container \"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a\": container with ID starting with 172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.631228 4960 scope.go:117] "RemoveContainer" containerID="022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554" Oct 02 07:34:24 crc kubenswrapper[4960]: E1002 07:34:24.631667 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554\": container with ID starting with 022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554 not found: ID does not exist" containerID="022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.631691 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554"} err="failed to get container status \"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554\": rpc error: code = NotFound desc = could not find container \"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554\": container with ID starting with 022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554 not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.631704 4960 scope.go:117] "RemoveContainer" containerID="bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.632006 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3"} err="failed to get container status \"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3\": rpc error: code = NotFound desc = could not find container \"bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3\": container with ID starting with bd4c235ee58904bbc230b27fceb540f24ded7568c83ddb5159d656e5844b90d3 not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.632043 4960 scope.go:117] "RemoveContainer" containerID="40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.632280 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db"} err="failed to get container status \"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db\": rpc error: code = NotFound desc = could not find container \"40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db\": container with ID starting with 40f366146b40bb52d23c307e15547d4e0ba957dc616c33b8c8b10cae004164db not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.632302 4960 scope.go:117] "RemoveContainer" containerID="172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.632500 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a"} err="failed to get container status \"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a\": rpc error: code = NotFound desc = could not find container \"172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a\": container with ID starting with 172a3276bc673c1b25ba52c472979763c436f18694d065a2b8e28a679ead7b0a not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.632520 4960 scope.go:117] "RemoveContainer" containerID="022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.632778 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554"} err="failed to get container status \"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554\": rpc error: code = NotFound desc = could not find container \"022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554\": container with ID starting with 022252295f0f74ce7debb04529edbaf94f609e77b5055d7273ef0b420f380554 not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.673774 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-log-httpd\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.674352 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.674486 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztxg7\" (UniqueName: \"kubernetes.io/projected/433e660d-96fa-4146-998d-dd2f36a64e32-kube-api-access-ztxg7\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.674573 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-run-httpd\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.674695 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-config-data\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.674780 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.674852 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-scripts\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.675060 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-log-httpd\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.675457 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-run-httpd\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.680833 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.685307 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-scripts\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.685721 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.686267 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-config-data\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.699922 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztxg7\" (UniqueName: \"kubernetes.io/projected/433e660d-96fa-4146-998d-dd2f36a64e32-kube-api-access-ztxg7\") pod \"ceilometer-0\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " pod="openstack/ceilometer-0" Oct 02 07:34:24 crc kubenswrapper[4960]: I1002 07:34:24.871252 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:25 crc kubenswrapper[4960]: I1002 07:34:25.075099 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 07:34:25 crc kubenswrapper[4960]: I1002 07:34:25.225872 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:25 crc kubenswrapper[4960]: I1002 07:34:25.343313 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:25 crc kubenswrapper[4960]: W1002 07:34:25.354807 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod433e660d_96fa_4146_998d_dd2f36a64e32.slice/crio-11f7abc7f30be8560600a51f7400f1ef865e3fbbbde334833335c5d5f7f2db45 WatchSource:0}: Error finding container 11f7abc7f30be8560600a51f7400f1ef865e3fbbbde334833335c5d5f7f2db45: Status 404 returned error can't find the container with id 11f7abc7f30be8560600a51f7400f1ef865e3fbbbde334833335c5d5f7f2db45 Oct 02 07:34:25 crc kubenswrapper[4960]: I1002 07:34:25.493697 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerStarted","Data":"11f7abc7f30be8560600a51f7400f1ef865e3fbbbde334833335c5d5f7f2db45"} Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.202885 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-2eca-account-create-4knbp"] Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.205562 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2eca-account-create-4knbp" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.209586 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.217330 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2eca-account-create-4knbp"] Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.313290 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4wz4\" (UniqueName: \"kubernetes.io/projected/c40986d9-621e-4cd4-abe0-c70c62baf39b-kube-api-access-b4wz4\") pod \"nova-api-2eca-account-create-4knbp\" (UID: \"c40986d9-621e-4cd4-abe0-c70c62baf39b\") " pod="openstack/nova-api-2eca-account-create-4knbp" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.341055 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efea504b-eb3a-41e9-804c-021e736dfd61" path="/var/lib/kubelet/pods/efea504b-eb3a-41e9-804c-021e736dfd61/volumes" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.404830 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-931c-account-create-t687z"] Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.406565 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-931c-account-create-t687z" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.408673 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.413294 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-931c-account-create-t687z"] Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.416721 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4wz4\" (UniqueName: \"kubernetes.io/projected/c40986d9-621e-4cd4-abe0-c70c62baf39b-kube-api-access-b4wz4\") pod \"nova-api-2eca-account-create-4knbp\" (UID: \"c40986d9-621e-4cd4-abe0-c70c62baf39b\") " pod="openstack/nova-api-2eca-account-create-4knbp" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.447707 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4wz4\" (UniqueName: \"kubernetes.io/projected/c40986d9-621e-4cd4-abe0-c70c62baf39b-kube-api-access-b4wz4\") pod \"nova-api-2eca-account-create-4knbp\" (UID: \"c40986d9-621e-4cd4-abe0-c70c62baf39b\") " pod="openstack/nova-api-2eca-account-create-4knbp" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.529229 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerStarted","Data":"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb"} Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.531799 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2eca-account-create-4knbp" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.533127 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xszhb\" (UniqueName: \"kubernetes.io/projected/6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f-kube-api-access-xszhb\") pod \"nova-cell0-931c-account-create-t687z\" (UID: \"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f\") " pod="openstack/nova-cell0-931c-account-create-t687z" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.650699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xszhb\" (UniqueName: \"kubernetes.io/projected/6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f-kube-api-access-xszhb\") pod \"nova-cell0-931c-account-create-t687z\" (UID: \"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f\") " pod="openstack/nova-cell0-931c-account-create-t687z" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.666423 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ca00-account-create-j24m7"] Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.669119 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca00-account-create-j24m7" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.672299 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.678693 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xszhb\" (UniqueName: \"kubernetes.io/projected/6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f-kube-api-access-xszhb\") pod \"nova-cell0-931c-account-create-t687z\" (UID: \"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f\") " pod="openstack/nova-cell0-931c-account-create-t687z" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.710196 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ca00-account-create-j24m7"] Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.737905 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-931c-account-create-t687z" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.757163 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr747\" (UniqueName: \"kubernetes.io/projected/2ad308ec-0722-4c1d-a871-062db558ec6b-kube-api-access-hr747\") pod \"nova-cell1-ca00-account-create-j24m7\" (UID: \"2ad308ec-0722-4c1d-a871-062db558ec6b\") " pod="openstack/nova-cell1-ca00-account-create-j24m7" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.858852 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr747\" (UniqueName: \"kubernetes.io/projected/2ad308ec-0722-4c1d-a871-062db558ec6b-kube-api-access-hr747\") pod \"nova-cell1-ca00-account-create-j24m7\" (UID: \"2ad308ec-0722-4c1d-a871-062db558ec6b\") " pod="openstack/nova-cell1-ca00-account-create-j24m7" Oct 02 07:34:26 crc kubenswrapper[4960]: I1002 07:34:26.886067 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr747\" (UniqueName: \"kubernetes.io/projected/2ad308ec-0722-4c1d-a871-062db558ec6b-kube-api-access-hr747\") pod \"nova-cell1-ca00-account-create-j24m7\" (UID: \"2ad308ec-0722-4c1d-a871-062db558ec6b\") " pod="openstack/nova-cell1-ca00-account-create-j24m7" Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.085882 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2eca-account-create-4knbp"] Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.100238 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca00-account-create-j24m7" Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.276812 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-931c-account-create-t687z"] Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.542497 4960 generic.go:334] "Generic (PLEG): container finished" podID="c40986d9-621e-4cd4-abe0-c70c62baf39b" containerID="a2f6419062ce4fadb106ed12c86a0b29065f7f8343c471da33d4a96508b0d751" exitCode=0 Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.543391 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2eca-account-create-4knbp" event={"ID":"c40986d9-621e-4cd4-abe0-c70c62baf39b","Type":"ContainerDied","Data":"a2f6419062ce4fadb106ed12c86a0b29065f7f8343c471da33d4a96508b0d751"} Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.543432 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2eca-account-create-4knbp" event={"ID":"c40986d9-621e-4cd4-abe0-c70c62baf39b","Type":"ContainerStarted","Data":"6b78832ee1bd285a87e338ac5618bc6b6f0c1b434658f530aa5bbf035541c810"} Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.545237 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-931c-account-create-t687z" event={"ID":"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f","Type":"ContainerStarted","Data":"33f3ad79cabd4d4be46ab0633313e8253f2c72f73b5090929180955636b941b6"} Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.545264 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-931c-account-create-t687z" event={"ID":"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f","Type":"ContainerStarted","Data":"3c9c07c266126a75321b4c755ac5e6e892a83a74dd3045ef257013d1092f5953"} Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.551374 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerStarted","Data":"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf"} Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.584357 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-931c-account-create-t687z" podStartSLOduration=1.5843330660000001 podStartE2EDuration="1.584333066s" podCreationTimestamp="2025-10-02 07:34:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:27.583280845 +0000 UTC m=+1088.615227132" watchObservedRunningTime="2025-10-02 07:34:27.584333066 +0000 UTC m=+1088.616279353" Oct 02 07:34:27 crc kubenswrapper[4960]: I1002 07:34:27.606903 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ca00-account-create-j24m7"] Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.241909 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.562476 4960 generic.go:334] "Generic (PLEG): container finished" podID="2ad308ec-0722-4c1d-a871-062db558ec6b" containerID="25f482babd0bd9deb5557faa195fa5669b05f0ec026db0ac52935e8c0c6e0b97" exitCode=0 Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.562535 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca00-account-create-j24m7" event={"ID":"2ad308ec-0722-4c1d-a871-062db558ec6b","Type":"ContainerDied","Data":"25f482babd0bd9deb5557faa195fa5669b05f0ec026db0ac52935e8c0c6e0b97"} Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.562564 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca00-account-create-j24m7" event={"ID":"2ad308ec-0722-4c1d-a871-062db558ec6b","Type":"ContainerStarted","Data":"9439c20333c8ed5f26d19ea31d35a04618244c765d623dda898d9dc816f18e27"} Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.565673 4960 generic.go:334] "Generic (PLEG): container finished" podID="6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f" containerID="33f3ad79cabd4d4be46ab0633313e8253f2c72f73b5090929180955636b941b6" exitCode=0 Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.565726 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-931c-account-create-t687z" event={"ID":"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f","Type":"ContainerDied","Data":"33f3ad79cabd4d4be46ab0633313e8253f2c72f73b5090929180955636b941b6"} Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.568568 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerStarted","Data":"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179"} Oct 02 07:34:28 crc kubenswrapper[4960]: I1002 07:34:28.987236 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2eca-account-create-4knbp" Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.109906 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4wz4\" (UniqueName: \"kubernetes.io/projected/c40986d9-621e-4cd4-abe0-c70c62baf39b-kube-api-access-b4wz4\") pod \"c40986d9-621e-4cd4-abe0-c70c62baf39b\" (UID: \"c40986d9-621e-4cd4-abe0-c70c62baf39b\") " Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.126040 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c40986d9-621e-4cd4-abe0-c70c62baf39b-kube-api-access-b4wz4" (OuterVolumeSpecName: "kube-api-access-b4wz4") pod "c40986d9-621e-4cd4-abe0-c70c62baf39b" (UID: "c40986d9-621e-4cd4-abe0-c70c62baf39b"). InnerVolumeSpecName "kube-api-access-b4wz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.212324 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4wz4\" (UniqueName: \"kubernetes.io/projected/c40986d9-621e-4cd4-abe0-c70c62baf39b-kube-api-access-b4wz4\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.582966 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerStarted","Data":"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b"} Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.583495 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-central-agent" containerID="cri-o://a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" gracePeriod=30 Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.583567 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.583605 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="proxy-httpd" containerID="cri-o://db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" gracePeriod=30 Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.583669 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="sg-core" containerID="cri-o://2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" gracePeriod=30 Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.583703 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-notification-agent" containerID="cri-o://908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" gracePeriod=30 Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.584950 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2eca-account-create-4knbp" Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.585316 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2eca-account-create-4knbp" event={"ID":"c40986d9-621e-4cd4-abe0-c70c62baf39b","Type":"ContainerDied","Data":"6b78832ee1bd285a87e338ac5618bc6b6f0c1b434658f530aa5bbf035541c810"} Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.585486 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b78832ee1bd285a87e338ac5618bc6b6f0c1b434658f530aa5bbf035541c810" Oct 02 07:34:29 crc kubenswrapper[4960]: I1002 07:34:29.624304 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.144575909 podStartE2EDuration="5.62428014s" podCreationTimestamp="2025-10-02 07:34:24 +0000 UTC" firstStartedPulling="2025-10-02 07:34:25.357660668 +0000 UTC m=+1086.389606955" lastFinishedPulling="2025-10-02 07:34:28.837364899 +0000 UTC m=+1089.869311186" observedRunningTime="2025-10-02 07:34:29.618768149 +0000 UTC m=+1090.650714446" watchObservedRunningTime="2025-10-02 07:34:29.62428014 +0000 UTC m=+1090.656226427" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.209172 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca00-account-create-j24m7" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.221145 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-931c-account-create-t687z" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.326685 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.336442 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr747\" (UniqueName: \"kubernetes.io/projected/2ad308ec-0722-4c1d-a871-062db558ec6b-kube-api-access-hr747\") pod \"2ad308ec-0722-4c1d-a871-062db558ec6b\" (UID: \"2ad308ec-0722-4c1d-a871-062db558ec6b\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.338615 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xszhb\" (UniqueName: \"kubernetes.io/projected/6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f-kube-api-access-xszhb\") pod \"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f\" (UID: \"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.343039 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ad308ec-0722-4c1d-a871-062db558ec6b-kube-api-access-hr747" (OuterVolumeSpecName: "kube-api-access-hr747") pod "2ad308ec-0722-4c1d-a871-062db558ec6b" (UID: "2ad308ec-0722-4c1d-a871-062db558ec6b"). InnerVolumeSpecName "kube-api-access-hr747". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.345229 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f-kube-api-access-xszhb" (OuterVolumeSpecName: "kube-api-access-xszhb") pod "6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f" (UID: "6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f"). InnerVolumeSpecName "kube-api-access-xszhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.416483 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.445259 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xszhb\" (UniqueName: \"kubernetes.io/projected/6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f-kube-api-access-xszhb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.445307 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr747\" (UniqueName: \"kubernetes.io/projected/2ad308ec-0722-4c1d-a871-062db558ec6b-kube-api-access-hr747\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.546773 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-sg-core-conf-yaml\") pod \"433e660d-96fa-4146-998d-dd2f36a64e32\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.546911 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-run-httpd\") pod \"433e660d-96fa-4146-998d-dd2f36a64e32\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.546951 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-log-httpd\") pod \"433e660d-96fa-4146-998d-dd2f36a64e32\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.547020 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-config-data\") pod \"433e660d-96fa-4146-998d-dd2f36a64e32\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.547114 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztxg7\" (UniqueName: \"kubernetes.io/projected/433e660d-96fa-4146-998d-dd2f36a64e32-kube-api-access-ztxg7\") pod \"433e660d-96fa-4146-998d-dd2f36a64e32\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.547163 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-scripts\") pod \"433e660d-96fa-4146-998d-dd2f36a64e32\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.547590 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "433e660d-96fa-4146-998d-dd2f36a64e32" (UID: "433e660d-96fa-4146-998d-dd2f36a64e32"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.547887 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "433e660d-96fa-4146-998d-dd2f36a64e32" (UID: "433e660d-96fa-4146-998d-dd2f36a64e32"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.547955 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-combined-ca-bundle\") pod \"433e660d-96fa-4146-998d-dd2f36a64e32\" (UID: \"433e660d-96fa-4146-998d-dd2f36a64e32\") " Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.548814 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.548835 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/433e660d-96fa-4146-998d-dd2f36a64e32-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.553084 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-scripts" (OuterVolumeSpecName: "scripts") pod "433e660d-96fa-4146-998d-dd2f36a64e32" (UID: "433e660d-96fa-4146-998d-dd2f36a64e32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.554557 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/433e660d-96fa-4146-998d-dd2f36a64e32-kube-api-access-ztxg7" (OuterVolumeSpecName: "kube-api-access-ztxg7") pod "433e660d-96fa-4146-998d-dd2f36a64e32" (UID: "433e660d-96fa-4146-998d-dd2f36a64e32"). InnerVolumeSpecName "kube-api-access-ztxg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.596708 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "433e660d-96fa-4146-998d-dd2f36a64e32" (UID: "433e660d-96fa-4146-998d-dd2f36a64e32"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601762 4960 generic.go:334] "Generic (PLEG): container finished" podID="433e660d-96fa-4146-998d-dd2f36a64e32" containerID="db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" exitCode=0 Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601794 4960 generic.go:334] "Generic (PLEG): container finished" podID="433e660d-96fa-4146-998d-dd2f36a64e32" containerID="2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" exitCode=2 Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601805 4960 generic.go:334] "Generic (PLEG): container finished" podID="433e660d-96fa-4146-998d-dd2f36a64e32" containerID="908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" exitCode=0 Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601818 4960 generic.go:334] "Generic (PLEG): container finished" podID="433e660d-96fa-4146-998d-dd2f36a64e32" containerID="a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" exitCode=0 Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601823 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601846 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerDied","Data":"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b"} Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601896 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerDied","Data":"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179"} Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601909 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerDied","Data":"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf"} Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601937 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerDied","Data":"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb"} Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601947 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"433e660d-96fa-4146-998d-dd2f36a64e32","Type":"ContainerDied","Data":"11f7abc7f30be8560600a51f7400f1ef865e3fbbbde334833335c5d5f7f2db45"} Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.601961 4960 scope.go:117] "RemoveContainer" containerID="db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.606140 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca00-account-create-j24m7" event={"ID":"2ad308ec-0722-4c1d-a871-062db558ec6b","Type":"ContainerDied","Data":"9439c20333c8ed5f26d19ea31d35a04618244c765d623dda898d9dc816f18e27"} Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.606185 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9439c20333c8ed5f26d19ea31d35a04618244c765d623dda898d9dc816f18e27" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.606247 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca00-account-create-j24m7" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.609708 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-931c-account-create-t687z" event={"ID":"6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f","Type":"ContainerDied","Data":"3c9c07c266126a75321b4c755ac5e6e892a83a74dd3045ef257013d1092f5953"} Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.609745 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c9c07c266126a75321b4c755ac5e6e892a83a74dd3045ef257013d1092f5953" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.609804 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-931c-account-create-t687z" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.638456 4960 scope.go:117] "RemoveContainer" containerID="2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.651051 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztxg7\" (UniqueName: \"kubernetes.io/projected/433e660d-96fa-4146-998d-dd2f36a64e32-kube-api-access-ztxg7\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.651098 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.651115 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.654862 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "433e660d-96fa-4146-998d-dd2f36a64e32" (UID: "433e660d-96fa-4146-998d-dd2f36a64e32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.661562 4960 scope.go:117] "RemoveContainer" containerID="908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.671684 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-config-data" (OuterVolumeSpecName: "config-data") pod "433e660d-96fa-4146-998d-dd2f36a64e32" (UID: "433e660d-96fa-4146-998d-dd2f36a64e32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.691035 4960 scope.go:117] "RemoveContainer" containerID="a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.711116 4960 scope.go:117] "RemoveContainer" containerID="db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.711816 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": container with ID starting with db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b not found: ID does not exist" containerID="db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.711865 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b"} err="failed to get container status \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": rpc error: code = NotFound desc = could not find container \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": container with ID starting with db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.711904 4960 scope.go:117] "RemoveContainer" containerID="2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.712453 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": container with ID starting with 2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179 not found: ID does not exist" containerID="2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.712511 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179"} err="failed to get container status \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": rpc error: code = NotFound desc = could not find container \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": container with ID starting with 2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179 not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.712561 4960 scope.go:117] "RemoveContainer" containerID="908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.713078 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": container with ID starting with 908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf not found: ID does not exist" containerID="908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.713134 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf"} err="failed to get container status \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": rpc error: code = NotFound desc = could not find container \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": container with ID starting with 908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.713173 4960 scope.go:117] "RemoveContainer" containerID="a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.713461 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": container with ID starting with a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb not found: ID does not exist" containerID="a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.713494 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb"} err="failed to get container status \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": rpc error: code = NotFound desc = could not find container \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": container with ID starting with a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.713515 4960 scope.go:117] "RemoveContainer" containerID="db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.713768 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b"} err="failed to get container status \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": rpc error: code = NotFound desc = could not find container \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": container with ID starting with db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.713833 4960 scope.go:117] "RemoveContainer" containerID="2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.714206 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179"} err="failed to get container status \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": rpc error: code = NotFound desc = could not find container \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": container with ID starting with 2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179 not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.714269 4960 scope.go:117] "RemoveContainer" containerID="908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.714640 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf"} err="failed to get container status \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": rpc error: code = NotFound desc = could not find container \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": container with ID starting with 908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.714690 4960 scope.go:117] "RemoveContainer" containerID="a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.714949 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb"} err="failed to get container status \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": rpc error: code = NotFound desc = could not find container \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": container with ID starting with a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.715019 4960 scope.go:117] "RemoveContainer" containerID="db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.715271 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b"} err="failed to get container status \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": rpc error: code = NotFound desc = could not find container \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": container with ID starting with db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.715299 4960 scope.go:117] "RemoveContainer" containerID="2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.715664 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179"} err="failed to get container status \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": rpc error: code = NotFound desc = could not find container \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": container with ID starting with 2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179 not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.715713 4960 scope.go:117] "RemoveContainer" containerID="908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.716008 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf"} err="failed to get container status \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": rpc error: code = NotFound desc = could not find container \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": container with ID starting with 908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.716032 4960 scope.go:117] "RemoveContainer" containerID="a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.717113 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb"} err="failed to get container status \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": rpc error: code = NotFound desc = could not find container \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": container with ID starting with a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.717143 4960 scope.go:117] "RemoveContainer" containerID="db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.717560 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b"} err="failed to get container status \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": rpc error: code = NotFound desc = could not find container \"db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b\": container with ID starting with db11915b2da229f63af98e2b3785caa4ff8e673bf00e148991fc3140420a8c6b not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.717590 4960 scope.go:117] "RemoveContainer" containerID="2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.717887 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179"} err="failed to get container status \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": rpc error: code = NotFound desc = could not find container \"2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179\": container with ID starting with 2b5b29518510a46ec2c9505515918f01c50b750a094833a5d237004275989179 not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.717920 4960 scope.go:117] "RemoveContainer" containerID="908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.718197 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf"} err="failed to get container status \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": rpc error: code = NotFound desc = could not find container \"908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf\": container with ID starting with 908807c3bb56d182bc92534287b267980ad8989b22158b90f111ae9028699fdf not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.718225 4960 scope.go:117] "RemoveContainer" containerID="a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.718546 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb"} err="failed to get container status \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": rpc error: code = NotFound desc = could not find container \"a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb\": container with ID starting with a862abb9493c987990de2fc7f7565de82efd7ded6d77d72e54e0bcf543042dcb not found: ID does not exist" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.755694 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.755818 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/433e660d-96fa-4146-998d-dd2f36a64e32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.955083 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.977784 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.988258 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.989538 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="proxy-httpd" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989554 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="proxy-httpd" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.989569 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40986d9-621e-4cd4-abe0-c70c62baf39b" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989578 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40986d9-621e-4cd4-abe0-c70c62baf39b" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.989600 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad308ec-0722-4c1d-a871-062db558ec6b" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989607 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad308ec-0722-4c1d-a871-062db558ec6b" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.989631 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-notification-agent" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989637 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-notification-agent" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.989652 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="sg-core" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989658 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="sg-core" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.989669 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989677 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: E1002 07:34:30.989694 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-central-agent" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989699 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-central-agent" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989864 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad308ec-0722-4c1d-a871-062db558ec6b" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989881 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="proxy-httpd" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989889 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-notification-agent" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989897 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="sg-core" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989909 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989922 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40986d9-621e-4cd4-abe0-c70c62baf39b" containerName="mariadb-account-create" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.989934 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" containerName="ceilometer-central-agent" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.991636 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:30 crc kubenswrapper[4960]: I1002 07:34:30.994273 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:30.999864 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.052640 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.163958 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-run-httpd\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.164017 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.164061 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-scripts\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.164082 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbtzc\" (UniqueName: \"kubernetes.io/projected/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-kube-api-access-kbtzc\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.164111 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-log-httpd\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.164159 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-config-data\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.164207 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.265627 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-scripts\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.265989 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbtzc\" (UniqueName: \"kubernetes.io/projected/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-kube-api-access-kbtzc\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.266109 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-log-httpd\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.266250 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-config-data\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.266421 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.266590 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-run-httpd\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.266699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.266699 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-log-httpd\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.266925 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-run-httpd\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.270246 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-scripts\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.270559 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-config-data\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.271952 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.272816 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.281303 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbtzc\" (UniqueName: \"kubernetes.io/projected/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-kube-api-access-kbtzc\") pod \"ceilometer-0\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.376629 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.589848 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-45zxh"] Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.591709 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.594331 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.595679 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mpqzv" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.596132 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.599745 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-45zxh"] Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.788722 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-config-data\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.788796 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qtkf\" (UniqueName: \"kubernetes.io/projected/b6ef2807-e246-4628-aa45-44ab0805d441-kube-api-access-4qtkf\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.788892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-scripts\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.788939 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.891132 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-config-data\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.891201 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qtkf\" (UniqueName: \"kubernetes.io/projected/b6ef2807-e246-4628-aa45-44ab0805d441-kube-api-access-4qtkf\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.891250 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-scripts\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.891284 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.901793 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-config-data\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.901869 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-scripts\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.904699 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.908942 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:31 crc kubenswrapper[4960]: I1002 07:34:31.926592 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qtkf\" (UniqueName: \"kubernetes.io/projected/b6ef2807-e246-4628-aa45-44ab0805d441-kube-api-access-4qtkf\") pod \"nova-cell0-conductor-db-sync-45zxh\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:32 crc kubenswrapper[4960]: I1002 07:34:32.222509 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:32 crc kubenswrapper[4960]: I1002 07:34:32.350697 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="433e660d-96fa-4146-998d-dd2f36a64e32" path="/var/lib/kubelet/pods/433e660d-96fa-4146-998d-dd2f36a64e32/volumes" Oct 02 07:34:32 crc kubenswrapper[4960]: I1002 07:34:32.643905 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerStarted","Data":"69274d46eca703db164e7cf1828eacdd29d787ccf789d9c4bbf4aedc0099ce51"} Oct 02 07:34:32 crc kubenswrapper[4960]: I1002 07:34:32.644311 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerStarted","Data":"83dcd1aef960028fe3e93a6f6c6f73cfcb7e9c60f1cbc009a111e2afc04ccd4b"} Oct 02 07:34:32 crc kubenswrapper[4960]: I1002 07:34:32.750476 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-45zxh"] Oct 02 07:34:33 crc kubenswrapper[4960]: I1002 07:34:33.657103 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-45zxh" event={"ID":"b6ef2807-e246-4628-aa45-44ab0805d441","Type":"ContainerStarted","Data":"b3154b121cadcd9b368054bd0a997b7ef6296c0c0b2cdd2a28e306a412e22eb8"} Oct 02 07:34:33 crc kubenswrapper[4960]: I1002 07:34:33.662207 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerStarted","Data":"0c301316575fd8d91b9cec70c6461e931fdfb73d734dbc22657941a660a5ca8c"} Oct 02 07:34:34 crc kubenswrapper[4960]: I1002 07:34:34.676087 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerStarted","Data":"520229791a049bb7abcc587234312d317ecf46519bc9294e51f53b56e0a26545"} Oct 02 07:34:35 crc kubenswrapper[4960]: I1002 07:34:35.711169 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerStarted","Data":"15cfb93839931bee4300e4777610d9a5e7202bf175bf2e09cd54d076b02432b8"} Oct 02 07:34:35 crc kubenswrapper[4960]: I1002 07:34:35.712443 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:34:38 crc kubenswrapper[4960]: I1002 07:34:38.658350 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.948967733 podStartE2EDuration="8.658326077s" podCreationTimestamp="2025-10-02 07:34:30 +0000 UTC" firstStartedPulling="2025-10-02 07:34:31.949398711 +0000 UTC m=+1092.981344998" lastFinishedPulling="2025-10-02 07:34:34.658757055 +0000 UTC m=+1095.690703342" observedRunningTime="2025-10-02 07:34:35.739302081 +0000 UTC m=+1096.771248388" watchObservedRunningTime="2025-10-02 07:34:38.658326077 +0000 UTC m=+1099.690272364" Oct 02 07:34:38 crc kubenswrapper[4960]: I1002 07:34:38.659092 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:38 crc kubenswrapper[4960]: I1002 07:34:38.740421 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-central-agent" containerID="cri-o://69274d46eca703db164e7cf1828eacdd29d787ccf789d9c4bbf4aedc0099ce51" gracePeriod=30 Oct 02 07:34:38 crc kubenswrapper[4960]: I1002 07:34:38.740527 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="sg-core" containerID="cri-o://520229791a049bb7abcc587234312d317ecf46519bc9294e51f53b56e0a26545" gracePeriod=30 Oct 02 07:34:38 crc kubenswrapper[4960]: I1002 07:34:38.740640 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-notification-agent" containerID="cri-o://0c301316575fd8d91b9cec70c6461e931fdfb73d734dbc22657941a660a5ca8c" gracePeriod=30 Oct 02 07:34:38 crc kubenswrapper[4960]: I1002 07:34:38.740697 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="proxy-httpd" containerID="cri-o://15cfb93839931bee4300e4777610d9a5e7202bf175bf2e09cd54d076b02432b8" gracePeriod=30 Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.759714 4960 generic.go:334] "Generic (PLEG): container finished" podID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerID="15cfb93839931bee4300e4777610d9a5e7202bf175bf2e09cd54d076b02432b8" exitCode=0 Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.760298 4960 generic.go:334] "Generic (PLEG): container finished" podID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerID="520229791a049bb7abcc587234312d317ecf46519bc9294e51f53b56e0a26545" exitCode=2 Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.759806 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerDied","Data":"15cfb93839931bee4300e4777610d9a5e7202bf175bf2e09cd54d076b02432b8"} Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.760363 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerDied","Data":"520229791a049bb7abcc587234312d317ecf46519bc9294e51f53b56e0a26545"} Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.760380 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerDied","Data":"0c301316575fd8d91b9cec70c6461e931fdfb73d734dbc22657941a660a5ca8c"} Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.760315 4960 generic.go:334] "Generic (PLEG): container finished" podID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerID="0c301316575fd8d91b9cec70c6461e931fdfb73d734dbc22657941a660a5ca8c" exitCode=0 Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.760410 4960 generic.go:334] "Generic (PLEG): container finished" podID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerID="69274d46eca703db164e7cf1828eacdd29d787ccf789d9c4bbf4aedc0099ce51" exitCode=0 Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.760436 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerDied","Data":"69274d46eca703db164e7cf1828eacdd29d787ccf789d9c4bbf4aedc0099ce51"} Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.795290 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.972834 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-combined-ca-bundle\") pod \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.973653 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-log-httpd\") pod \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.974189 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-config-data\") pod \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.974289 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-scripts\") pod \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.974386 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-sg-core-conf-yaml\") pod \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.974472 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-run-httpd\") pod \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.974588 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" (UID: "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.974736 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbtzc\" (UniqueName: \"kubernetes.io/projected/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-kube-api-access-kbtzc\") pod \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\" (UID: \"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66\") " Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.974887 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" (UID: "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.976013 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.976116 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.980585 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-kube-api-access-kbtzc" (OuterVolumeSpecName: "kube-api-access-kbtzc") pod "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" (UID: "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66"). InnerVolumeSpecName "kube-api-access-kbtzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:39 crc kubenswrapper[4960]: I1002 07:34:39.980952 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-scripts" (OuterVolumeSpecName: "scripts") pod "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" (UID: "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.013080 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" (UID: "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.049707 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" (UID: "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.079050 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbtzc\" (UniqueName: \"kubernetes.io/projected/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-kube-api-access-kbtzc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.079092 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.079105 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.079116 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.091476 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-config-data" (OuterVolumeSpecName: "config-data") pod "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" (UID: "4e8bf592-4fdf-4f48-ad95-5c87ae16ee66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.181140 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.774719 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-45zxh" event={"ID":"b6ef2807-e246-4628-aa45-44ab0805d441","Type":"ContainerStarted","Data":"6da09db331c50808dd892510883439a1650cbd3d3e4951a9521f62a22380f5a0"} Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.779431 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e8bf592-4fdf-4f48-ad95-5c87ae16ee66","Type":"ContainerDied","Data":"83dcd1aef960028fe3e93a6f6c6f73cfcb7e9c60f1cbc009a111e2afc04ccd4b"} Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.779496 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.779522 4960 scope.go:117] "RemoveContainer" containerID="15cfb93839931bee4300e4777610d9a5e7202bf175bf2e09cd54d076b02432b8" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.806093 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-45zxh" podStartSLOduration=3.020911062 podStartE2EDuration="9.806025123s" podCreationTimestamp="2025-10-02 07:34:31 +0000 UTC" firstStartedPulling="2025-10-02 07:34:32.756036168 +0000 UTC m=+1093.787982455" lastFinishedPulling="2025-10-02 07:34:39.541150209 +0000 UTC m=+1100.573096516" observedRunningTime="2025-10-02 07:34:40.790949264 +0000 UTC m=+1101.822895571" watchObservedRunningTime="2025-10-02 07:34:40.806025123 +0000 UTC m=+1101.837971440" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.830962 4960 scope.go:117] "RemoveContainer" containerID="520229791a049bb7abcc587234312d317ecf46519bc9294e51f53b56e0a26545" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.847218 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.856397 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.866836 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:40 crc kubenswrapper[4960]: E1002 07:34:40.867661 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-central-agent" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.867692 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-central-agent" Oct 02 07:34:40 crc kubenswrapper[4960]: E1002 07:34:40.867726 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="sg-core" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.867735 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="sg-core" Oct 02 07:34:40 crc kubenswrapper[4960]: E1002 07:34:40.867754 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="proxy-httpd" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.867763 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="proxy-httpd" Oct 02 07:34:40 crc kubenswrapper[4960]: E1002 07:34:40.867776 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-notification-agent" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.867784 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-notification-agent" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.868114 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-central-agent" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.868139 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="ceilometer-notification-agent" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.868172 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="sg-core" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.868188 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" containerName="proxy-httpd" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.870858 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.874340 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.874436 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.876349 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.891276 4960 scope.go:117] "RemoveContainer" containerID="0c301316575fd8d91b9cec70c6461e931fdfb73d734dbc22657941a660a5ca8c" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.951888 4960 scope.go:117] "RemoveContainer" containerID="69274d46eca703db164e7cf1828eacdd29d787ccf789d9c4bbf4aedc0099ce51" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.997907 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-scripts\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.997956 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r5df\" (UniqueName: \"kubernetes.io/projected/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-kube-api-access-4r5df\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.997994 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-config-data\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.998038 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-log-httpd\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.998882 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-run-httpd\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.999180 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:40 crc kubenswrapper[4960]: I1002 07:34:40.999594 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.102134 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.102246 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-scripts\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.102280 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r5df\" (UniqueName: \"kubernetes.io/projected/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-kube-api-access-4r5df\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.102313 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-config-data\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.102363 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-log-httpd\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.102391 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-run-httpd\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.102458 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.104815 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-log-httpd\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.104851 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-run-httpd\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.107587 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.107918 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-scripts\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.110002 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-config-data\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.119056 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.127171 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r5df\" (UniqueName: \"kubernetes.io/projected/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-kube-api-access-4r5df\") pod \"ceilometer-0\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.216354 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.699926 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:41 crc kubenswrapper[4960]: I1002 07:34:41.793961 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerStarted","Data":"83ae0f48392ca3b18c5d8fe63403a5ff9b355dd70ffc8fe66c41ef801b180e25"} Oct 02 07:34:42 crc kubenswrapper[4960]: I1002 07:34:42.349896 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e8bf592-4fdf-4f48-ad95-5c87ae16ee66" path="/var/lib/kubelet/pods/4e8bf592-4fdf-4f48-ad95-5c87ae16ee66/volumes" Oct 02 07:34:42 crc kubenswrapper[4960]: I1002 07:34:42.817618 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerStarted","Data":"4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6"} Oct 02 07:34:43 crc kubenswrapper[4960]: I1002 07:34:43.837707 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerStarted","Data":"4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a"} Oct 02 07:34:46 crc kubenswrapper[4960]: I1002 07:34:46.876088 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerStarted","Data":"208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5"} Oct 02 07:34:47 crc kubenswrapper[4960]: I1002 07:34:47.894446 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerStarted","Data":"1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca"} Oct 02 07:34:47 crc kubenswrapper[4960]: I1002 07:34:47.897377 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:34:47 crc kubenswrapper[4960]: I1002 07:34:47.923455 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.054109017 podStartE2EDuration="7.923427338s" podCreationTimestamp="2025-10-02 07:34:40 +0000 UTC" firstStartedPulling="2025-10-02 07:34:41.709048878 +0000 UTC m=+1102.740995195" lastFinishedPulling="2025-10-02 07:34:47.578367229 +0000 UTC m=+1108.610313516" observedRunningTime="2025-10-02 07:34:47.920205861 +0000 UTC m=+1108.952152178" watchObservedRunningTime="2025-10-02 07:34:47.923427338 +0000 UTC m=+1108.955373625" Oct 02 07:34:48 crc kubenswrapper[4960]: I1002 07:34:48.501087 4960 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod19b6c4e4-068d-4969-a8b3-95d907bde1c5"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod19b6c4e4-068d-4969-a8b3-95d907bde1c5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod19b6c4e4_068d_4969_a8b3_95d907bde1c5.slice" Oct 02 07:34:48 crc kubenswrapper[4960]: E1002 07:34:48.501178 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod19b6c4e4-068d-4969-a8b3-95d907bde1c5] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod19b6c4e4-068d-4969-a8b3-95d907bde1c5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod19b6c4e4_068d_4969_a8b3_95d907bde1c5.slice" pod="openstack/neutron-6ff76df558-2w955" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" Oct 02 07:34:48 crc kubenswrapper[4960]: I1002 07:34:48.908221 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ff76df558-2w955" Oct 02 07:34:48 crc kubenswrapper[4960]: I1002 07:34:48.979878 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ff76df558-2w955"] Oct 02 07:34:48 crc kubenswrapper[4960]: I1002 07:34:48.986959 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6ff76df558-2w955"] Oct 02 07:34:50 crc kubenswrapper[4960]: I1002 07:34:50.347131 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19b6c4e4-068d-4969-a8b3-95d907bde1c5" path="/var/lib/kubelet/pods/19b6c4e4-068d-4969-a8b3-95d907bde1c5/volumes" Oct 02 07:34:50 crc kubenswrapper[4960]: I1002 07:34:50.937243 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6ef2807-e246-4628-aa45-44ab0805d441" containerID="6da09db331c50808dd892510883439a1650cbd3d3e4951a9521f62a22380f5a0" exitCode=0 Oct 02 07:34:50 crc kubenswrapper[4960]: I1002 07:34:50.937604 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-45zxh" event={"ID":"b6ef2807-e246-4628-aa45-44ab0805d441","Type":"ContainerDied","Data":"6da09db331c50808dd892510883439a1650cbd3d3e4951a9521f62a22380f5a0"} Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.330811 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.365255 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-scripts\") pod \"b6ef2807-e246-4628-aa45-44ab0805d441\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.365333 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-combined-ca-bundle\") pod \"b6ef2807-e246-4628-aa45-44ab0805d441\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.365398 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-config-data\") pod \"b6ef2807-e246-4628-aa45-44ab0805d441\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.365446 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qtkf\" (UniqueName: \"kubernetes.io/projected/b6ef2807-e246-4628-aa45-44ab0805d441-kube-api-access-4qtkf\") pod \"b6ef2807-e246-4628-aa45-44ab0805d441\" (UID: \"b6ef2807-e246-4628-aa45-44ab0805d441\") " Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.374195 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-scripts" (OuterVolumeSpecName: "scripts") pod "b6ef2807-e246-4628-aa45-44ab0805d441" (UID: "b6ef2807-e246-4628-aa45-44ab0805d441"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.383751 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6ef2807-e246-4628-aa45-44ab0805d441-kube-api-access-4qtkf" (OuterVolumeSpecName: "kube-api-access-4qtkf") pod "b6ef2807-e246-4628-aa45-44ab0805d441" (UID: "b6ef2807-e246-4628-aa45-44ab0805d441"). InnerVolumeSpecName "kube-api-access-4qtkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.402454 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-config-data" (OuterVolumeSpecName: "config-data") pod "b6ef2807-e246-4628-aa45-44ab0805d441" (UID: "b6ef2807-e246-4628-aa45-44ab0805d441"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.423455 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6ef2807-e246-4628-aa45-44ab0805d441" (UID: "b6ef2807-e246-4628-aa45-44ab0805d441"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.468033 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qtkf\" (UniqueName: \"kubernetes.io/projected/b6ef2807-e246-4628-aa45-44ab0805d441-kube-api-access-4qtkf\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.468214 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.468238 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.468258 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ef2807-e246-4628-aa45-44ab0805d441-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.964755 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-45zxh" event={"ID":"b6ef2807-e246-4628-aa45-44ab0805d441","Type":"ContainerDied","Data":"b3154b121cadcd9b368054bd0a997b7ef6296c0c0b2cdd2a28e306a412e22eb8"} Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.964858 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3154b121cadcd9b368054bd0a997b7ef6296c0c0b2cdd2a28e306a412e22eb8" Oct 02 07:34:52 crc kubenswrapper[4960]: I1002 07:34:52.964910 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-45zxh" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.187597 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 07:34:53 crc kubenswrapper[4960]: E1002 07:34:53.188357 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6ef2807-e246-4628-aa45-44ab0805d441" containerName="nova-cell0-conductor-db-sync" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.188391 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6ef2807-e246-4628-aa45-44ab0805d441" containerName="nova-cell0-conductor-db-sync" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.188761 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6ef2807-e246-4628-aa45-44ab0805d441" containerName="nova-cell0-conductor-db-sync" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.190280 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.194937 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mpqzv" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.197081 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.209154 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.388345 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.388908 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.389039 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvkzx\" (UniqueName: \"kubernetes.io/projected/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-kube-api-access-zvkzx\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.491886 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvkzx\" (UniqueName: \"kubernetes.io/projected/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-kube-api-access-zvkzx\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.492553 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.493877 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.498311 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.500337 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.524270 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvkzx\" (UniqueName: \"kubernetes.io/projected/e82f1c13-1224-4c25-9050-d55a2ff7c5e6-kube-api-access-zvkzx\") pod \"nova-cell0-conductor-0\" (UID: \"e82f1c13-1224-4c25-9050-d55a2ff7c5e6\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:53 crc kubenswrapper[4960]: I1002 07:34:53.812729 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:54 crc kubenswrapper[4960]: I1002 07:34:54.366886 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 07:34:55 crc kubenswrapper[4960]: I1002 07:34:55.007296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e82f1c13-1224-4c25-9050-d55a2ff7c5e6","Type":"ContainerStarted","Data":"31562c7a11bbac3096c970f1458e0b4efe189077a5d12b739b8258be5cf4c470"} Oct 02 07:34:55 crc kubenswrapper[4960]: I1002 07:34:55.007784 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e82f1c13-1224-4c25-9050-d55a2ff7c5e6","Type":"ContainerStarted","Data":"d10f9df0dd840bc385a9993f178ea1eb773d6da5933f9ed86dd7ed159be95c53"} Oct 02 07:34:55 crc kubenswrapper[4960]: I1002 07:34:55.008164 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 07:34:55 crc kubenswrapper[4960]: I1002 07:34:55.046808 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.046782936 podStartE2EDuration="2.046782936s" podCreationTimestamp="2025-10-02 07:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:55.034950225 +0000 UTC m=+1116.066896542" watchObservedRunningTime="2025-10-02 07:34:55.046782936 +0000 UTC m=+1116.078729233" Oct 02 07:34:59 crc kubenswrapper[4960]: I1002 07:34:59.149785 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:34:59 crc kubenswrapper[4960]: I1002 07:34:59.150753 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:35:03 crc kubenswrapper[4960]: I1002 07:35:03.862715 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.403756 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-7nd6x"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.406187 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.410192 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.410582 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.415871 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7nd6x"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.562353 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t5p7\" (UniqueName: \"kubernetes.io/projected/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-kube-api-access-5t5p7\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.562507 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-config-data\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.562567 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-scripts\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.562591 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.623778 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.632309 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.636376 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.646008 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.664926 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-scripts\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.665035 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.666622 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t5p7\" (UniqueName: \"kubernetes.io/projected/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-kube-api-access-5t5p7\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.666807 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-config-data\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.673096 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-scripts\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.675911 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-config-data\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.679963 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.715885 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t5p7\" (UniqueName: \"kubernetes.io/projected/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-kube-api-access-5t5p7\") pod \"nova-cell0-cell-mapping-7nd6x\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.740656 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.766301 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.767792 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.768403 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5n8x\" (UniqueName: \"kubernetes.io/projected/37ce726a-c9d7-46af-840f-9df5f9266721-kube-api-access-t5n8x\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.768469 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37ce726a-c9d7-46af-840f-9df5f9266721-logs\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.768516 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.768607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-config-data\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.776416 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.822404 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.824210 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.826849 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.862760 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.872307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqwl6\" (UniqueName: \"kubernetes.io/projected/9d6b840e-7ada-4631-b63c-6d1edcf42b21-kube-api-access-kqwl6\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.872378 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.872401 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-config-data\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.872431 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5n8x\" (UniqueName: \"kubernetes.io/projected/37ce726a-c9d7-46af-840f-9df5f9266721-kube-api-access-t5n8x\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.872471 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37ce726a-c9d7-46af-840f-9df5f9266721-logs\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.872505 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.872567 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-config-data\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.876670 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37ce726a-c9d7-46af-840f-9df5f9266721-logs\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.886552 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-config-data\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.900053 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.915866 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.924749 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5n8x\" (UniqueName: \"kubernetes.io/projected/37ce726a-c9d7-46af-840f-9df5f9266721-kube-api-access-t5n8x\") pod \"nova-api-0\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.959023 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.980116 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-996hm\" (UniqueName: \"kubernetes.io/projected/f147cb3a-614c-4196-8c40-d1f974dc82a0-kube-api-access-996hm\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.982041 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqwl6\" (UniqueName: \"kubernetes.io/projected/9d6b840e-7ada-4631-b63c-6d1edcf42b21-kube-api-access-kqwl6\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.982152 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-config-data\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.982333 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.982446 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.982542 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-config-data\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.982710 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f147cb3a-614c-4196-8c40-d1f974dc82a0-logs\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.990648 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-config-data\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.990823 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.996495 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69775dfb85-5zwxs"] Oct 02 07:35:04 crc kubenswrapper[4960]: I1002 07:35:04.998131 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.031124 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqwl6\" (UniqueName: \"kubernetes.io/projected/9d6b840e-7ada-4631-b63c-6d1edcf42b21-kube-api-access-kqwl6\") pod \"nova-scheduler-0\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.042898 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.046282 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.050565 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.054528 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69775dfb85-5zwxs"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.069920 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.084937 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f147cb3a-614c-4196-8c40-d1f974dc82a0-logs\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.085002 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-nb\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.085081 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-996hm\" (UniqueName: \"kubernetes.io/projected/f147cb3a-614c-4196-8c40-d1f974dc82a0-kube-api-access-996hm\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.085136 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-config\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.085164 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-dns-svc\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.085183 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-config-data\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.085209 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48sjw\" (UniqueName: \"kubernetes.io/projected/fc70035e-3a13-476f-a5de-e13255cd4a5a-kube-api-access-48sjw\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.085256 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.088416 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f147cb3a-614c-4196-8c40-d1f974dc82a0-logs\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.091145 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-sb\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.096998 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-config-data\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.098280 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.108812 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-996hm\" (UniqueName: \"kubernetes.io/projected/f147cb3a-614c-4196-8c40-d1f974dc82a0-kube-api-access-996hm\") pod \"nova-metadata-0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.197754 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.197860 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-sb\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.197904 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-nb\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.197936 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.198087 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-config\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.198116 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-dns-svc\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.198142 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48sjw\" (UniqueName: \"kubernetes.io/projected/fc70035e-3a13-476f-a5de-e13255cd4a5a-kube-api-access-48sjw\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.198199 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5x8f\" (UniqueName: \"kubernetes.io/projected/0981f94f-b2f4-46dd-a7c0-f6b761960f30-kube-api-access-w5x8f\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.199186 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-nb\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.199243 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-config\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.199374 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-sb\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.199789 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-dns-svc\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.230520 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48sjw\" (UniqueName: \"kubernetes.io/projected/fc70035e-3a13-476f-a5de-e13255cd4a5a-kube-api-access-48sjw\") pod \"dnsmasq-dns-69775dfb85-5zwxs\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.264162 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.281346 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.299786 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5x8f\" (UniqueName: \"kubernetes.io/projected/0981f94f-b2f4-46dd-a7c0-f6b761960f30-kube-api-access-w5x8f\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.299848 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.299923 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.306840 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.309600 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.321764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5x8f\" (UniqueName: \"kubernetes.io/projected/0981f94f-b2f4-46dd-a7c0-f6b761960f30-kube-api-access-w5x8f\") pod \"nova-cell1-novncproxy-0\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.349031 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.391661 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.522638 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7nd6x"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.555347 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.586889 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zxzpz"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.589104 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.594561 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.594563 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.621603 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zxzpz"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.722797 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-scripts\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.723508 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.723541 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-config-data\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.723600 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc87b\" (UniqueName: \"kubernetes.io/projected/334bd1b9-aae9-4744-b800-817bbf01d3d9-kube-api-access-zc87b\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.827762 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-scripts\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.827843 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.827912 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-config-data\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.827957 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc87b\" (UniqueName: \"kubernetes.io/projected/334bd1b9-aae9-4744-b800-817bbf01d3d9-kube-api-access-zc87b\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.834541 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-scripts\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.846833 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.848616 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-config-data\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.850992 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc87b\" (UniqueName: \"kubernetes.io/projected/334bd1b9-aae9-4744-b800-817bbf01d3d9-kube-api-access-zc87b\") pod \"nova-cell1-conductor-db-sync-zxzpz\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.893943 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:05 crc kubenswrapper[4960]: I1002 07:35:05.928469 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.015192 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69775dfb85-5zwxs"] Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.039308 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.174238 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" event={"ID":"fc70035e-3a13-476f-a5de-e13255cd4a5a","Type":"ContainerStarted","Data":"28671a7eca61b82f9cf5f42cd3bde6c2a3a7e1ac7046c0bf35fb67afd84d8adc"} Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.190775 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f147cb3a-614c-4196-8c40-d1f974dc82a0","Type":"ContainerStarted","Data":"f78964bdae0cd3a983276b10b68d2a939e4ce0c08eb7b2553da77fd725e4f9ad"} Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.196551 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d6b840e-7ada-4631-b63c-6d1edcf42b21","Type":"ContainerStarted","Data":"11a97680f7ecd2d7ba8b8f9eb05ad47d236a9a32a905dfae6a0cac893deb697c"} Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.203923 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:06 crc kubenswrapper[4960]: W1002 07:35:06.210223 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0981f94f_b2f4_46dd_a7c0_f6b761960f30.slice/crio-b6e719d5cedd084555a2f721645fe71b21984dba02809ddc708346af5f141f94 WatchSource:0}: Error finding container b6e719d5cedd084555a2f721645fe71b21984dba02809ddc708346af5f141f94: Status 404 returned error can't find the container with id b6e719d5cedd084555a2f721645fe71b21984dba02809ddc708346af5f141f94 Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.213034 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37ce726a-c9d7-46af-840f-9df5f9266721","Type":"ContainerStarted","Data":"182279efd55acf48769e07caeeb2edfcfb583fea19b27fd69983b7a941d3ac10"} Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.216885 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7nd6x" event={"ID":"b8106fcd-0e3e-4440-bbc7-96cd83d369b5","Type":"ContainerStarted","Data":"7a94229645b189377e83e8ffd739528a27128b8b300793c4fb16791c23e9cf81"} Oct 02 07:35:06 crc kubenswrapper[4960]: I1002 07:35:06.481510 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zxzpz"] Oct 02 07:35:06 crc kubenswrapper[4960]: W1002 07:35:06.489327 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod334bd1b9_aae9_4744_b800_817bbf01d3d9.slice/crio-d399d7729dcd45522136e836b6c3a052d3cb4fdce84fdd3f83a325f73870cfac WatchSource:0}: Error finding container d399d7729dcd45522136e836b6c3a052d3cb4fdce84fdd3f83a325f73870cfac: Status 404 returned error can't find the container with id d399d7729dcd45522136e836b6c3a052d3cb4fdce84fdd3f83a325f73870cfac Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.235504 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7nd6x" event={"ID":"b8106fcd-0e3e-4440-bbc7-96cd83d369b5","Type":"ContainerStarted","Data":"6389c63ca445a838d188e1446238f1d73f34593d1ba53928214879d4167919c4"} Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.244265 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0981f94f-b2f4-46dd-a7c0-f6b761960f30","Type":"ContainerStarted","Data":"b6e719d5cedd084555a2f721645fe71b21984dba02809ddc708346af5f141f94"} Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.268173 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-7nd6x" podStartSLOduration=3.268146771 podStartE2EDuration="3.268146771s" podCreationTimestamp="2025-10-02 07:35:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:07.26405497 +0000 UTC m=+1128.296001267" watchObservedRunningTime="2025-10-02 07:35:07.268146771 +0000 UTC m=+1128.300093058" Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.268942 4960 generic.go:334] "Generic (PLEG): container finished" podID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerID="2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89" exitCode=0 Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.269049 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" event={"ID":"fc70035e-3a13-476f-a5de-e13255cd4a5a","Type":"ContainerDied","Data":"2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89"} Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.274451 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" event={"ID":"334bd1b9-aae9-4744-b800-817bbf01d3d9","Type":"ContainerStarted","Data":"cfcb3e868db40531716a774898dfaab314ca1fc0e1bf24fef47df76f39fcf6d0"} Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.274512 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" event={"ID":"334bd1b9-aae9-4744-b800-817bbf01d3d9","Type":"ContainerStarted","Data":"d399d7729dcd45522136e836b6c3a052d3cb4fdce84fdd3f83a325f73870cfac"} Oct 02 07:35:07 crc kubenswrapper[4960]: I1002 07:35:07.331523 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" podStartSLOduration=2.33149727 podStartE2EDuration="2.33149727s" podCreationTimestamp="2025-10-02 07:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:07.323934624 +0000 UTC m=+1128.355880911" watchObservedRunningTime="2025-10-02 07:35:07.33149727 +0000 UTC m=+1128.363443557" Oct 02 07:35:08 crc kubenswrapper[4960]: I1002 07:35:08.516262 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:08 crc kubenswrapper[4960]: I1002 07:35:08.554064 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.225464 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.332830 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" event={"ID":"fc70035e-3a13-476f-a5de-e13255cd4a5a","Type":"ContainerStarted","Data":"b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a"} Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.333029 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.338725 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f147cb3a-614c-4196-8c40-d1f974dc82a0","Type":"ContainerStarted","Data":"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d"} Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.338775 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f147cb3a-614c-4196-8c40-d1f974dc82a0","Type":"ContainerStarted","Data":"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8"} Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.338931 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-log" containerID="cri-o://52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8" gracePeriod=30 Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.339119 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-metadata" containerID="cri-o://cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d" gracePeriod=30 Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.345816 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d6b840e-7ada-4631-b63c-6d1edcf42b21","Type":"ContainerStarted","Data":"bb013ef65473a5c4ee434838fd66162c8629a9a9ea573902b16eaae58f971dcf"} Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.351362 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37ce726a-c9d7-46af-840f-9df5f9266721","Type":"ContainerStarted","Data":"67c2b2c207b3199ffbe81b5e678f0afca32b8670c354e801d89153f19801fdd5"} Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.351409 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37ce726a-c9d7-46af-840f-9df5f9266721","Type":"ContainerStarted","Data":"0218a1adbe98c03440bf9db253b26bed017c490bc0e7056af40abc9326f02987"} Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.362294 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0981f94f-b2f4-46dd-a7c0-f6b761960f30","Type":"ContainerStarted","Data":"c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801"} Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.362734 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0981f94f-b2f4-46dd-a7c0-f6b761960f30" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801" gracePeriod=30 Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.374895 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" podStartSLOduration=7.374870979 podStartE2EDuration="7.374870979s" podCreationTimestamp="2025-10-02 07:35:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:11.357171129 +0000 UTC m=+1132.389117416" watchObservedRunningTime="2025-10-02 07:35:11.374870979 +0000 UTC m=+1132.406817266" Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.382312 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.065569986 podStartE2EDuration="7.382289991s" podCreationTimestamp="2025-10-02 07:35:04 +0000 UTC" firstStartedPulling="2025-10-02 07:35:05.909745533 +0000 UTC m=+1126.941691820" lastFinishedPulling="2025-10-02 07:35:10.226465508 +0000 UTC m=+1131.258411825" observedRunningTime="2025-10-02 07:35:11.381392956 +0000 UTC m=+1132.413339263" watchObservedRunningTime="2025-10-02 07:35:11.382289991 +0000 UTC m=+1132.414236278" Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.407933 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.813852668 podStartE2EDuration="7.407912726s" podCreationTimestamp="2025-10-02 07:35:04 +0000 UTC" firstStartedPulling="2025-10-02 07:35:05.595833568 +0000 UTC m=+1126.627779855" lastFinishedPulling="2025-10-02 07:35:10.189893626 +0000 UTC m=+1131.221839913" observedRunningTime="2025-10-02 07:35:11.402465488 +0000 UTC m=+1132.434411775" watchObservedRunningTime="2025-10-02 07:35:11.407912726 +0000 UTC m=+1132.439859013" Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.439952 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.28506126 podStartE2EDuration="7.439921724s" podCreationTimestamp="2025-10-02 07:35:04 +0000 UTC" firstStartedPulling="2025-10-02 07:35:06.072310843 +0000 UTC m=+1127.104257130" lastFinishedPulling="2025-10-02 07:35:10.227171307 +0000 UTC m=+1131.259117594" observedRunningTime="2025-10-02 07:35:11.425182794 +0000 UTC m=+1132.457129091" watchObservedRunningTime="2025-10-02 07:35:11.439921724 +0000 UTC m=+1132.471868011" Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.471272 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.449534671 podStartE2EDuration="7.471245273s" podCreationTimestamp="2025-10-02 07:35:04 +0000 UTC" firstStartedPulling="2025-10-02 07:35:06.213106822 +0000 UTC m=+1127.245053109" lastFinishedPulling="2025-10-02 07:35:10.234817384 +0000 UTC m=+1131.266763711" observedRunningTime="2025-10-02 07:35:11.463678439 +0000 UTC m=+1132.495624726" watchObservedRunningTime="2025-10-02 07:35:11.471245273 +0000 UTC m=+1132.503191560" Oct 02 07:35:11 crc kubenswrapper[4960]: I1002 07:35:11.975619 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.093386 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-996hm\" (UniqueName: \"kubernetes.io/projected/f147cb3a-614c-4196-8c40-d1f974dc82a0-kube-api-access-996hm\") pod \"f147cb3a-614c-4196-8c40-d1f974dc82a0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.093499 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f147cb3a-614c-4196-8c40-d1f974dc82a0-logs\") pod \"f147cb3a-614c-4196-8c40-d1f974dc82a0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.093546 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-config-data\") pod \"f147cb3a-614c-4196-8c40-d1f974dc82a0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.093734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-combined-ca-bundle\") pod \"f147cb3a-614c-4196-8c40-d1f974dc82a0\" (UID: \"f147cb3a-614c-4196-8c40-d1f974dc82a0\") " Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.094601 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f147cb3a-614c-4196-8c40-d1f974dc82a0-logs" (OuterVolumeSpecName: "logs") pod "f147cb3a-614c-4196-8c40-d1f974dc82a0" (UID: "f147cb3a-614c-4196-8c40-d1f974dc82a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.101093 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f147cb3a-614c-4196-8c40-d1f974dc82a0-kube-api-access-996hm" (OuterVolumeSpecName: "kube-api-access-996hm") pod "f147cb3a-614c-4196-8c40-d1f974dc82a0" (UID: "f147cb3a-614c-4196-8c40-d1f974dc82a0"). InnerVolumeSpecName "kube-api-access-996hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.128151 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f147cb3a-614c-4196-8c40-d1f974dc82a0" (UID: "f147cb3a-614c-4196-8c40-d1f974dc82a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.136288 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-config-data" (OuterVolumeSpecName: "config-data") pod "f147cb3a-614c-4196-8c40-d1f974dc82a0" (UID: "f147cb3a-614c-4196-8c40-d1f974dc82a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.196662 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-996hm\" (UniqueName: \"kubernetes.io/projected/f147cb3a-614c-4196-8c40-d1f974dc82a0-kube-api-access-996hm\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.196714 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f147cb3a-614c-4196-8c40-d1f974dc82a0-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.196729 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.196739 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f147cb3a-614c-4196-8c40-d1f974dc82a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.377894 4960 generic.go:334] "Generic (PLEG): container finished" podID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerID="cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d" exitCode=0 Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.377946 4960 generic.go:334] "Generic (PLEG): container finished" podID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerID="52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8" exitCode=143 Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.378192 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f147cb3a-614c-4196-8c40-d1f974dc82a0","Type":"ContainerDied","Data":"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d"} Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.378296 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.378330 4960 scope.go:117] "RemoveContainer" containerID="cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.378308 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f147cb3a-614c-4196-8c40-d1f974dc82a0","Type":"ContainerDied","Data":"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8"} Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.378441 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f147cb3a-614c-4196-8c40-d1f974dc82a0","Type":"ContainerDied","Data":"f78964bdae0cd3a983276b10b68d2a939e4ce0c08eb7b2553da77fd725e4f9ad"} Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.416151 4960 scope.go:117] "RemoveContainer" containerID="52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.435594 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.441233 4960 scope.go:117] "RemoveContainer" containerID="cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d" Oct 02 07:35:12 crc kubenswrapper[4960]: E1002 07:35:12.442827 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d\": container with ID starting with cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d not found: ID does not exist" containerID="cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.442864 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d"} err="failed to get container status \"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d\": rpc error: code = NotFound desc = could not find container \"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d\": container with ID starting with cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d not found: ID does not exist" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.442904 4960 scope.go:117] "RemoveContainer" containerID="52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8" Oct 02 07:35:12 crc kubenswrapper[4960]: E1002 07:35:12.444247 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8\": container with ID starting with 52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8 not found: ID does not exist" containerID="52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.444349 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8"} err="failed to get container status \"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8\": rpc error: code = NotFound desc = could not find container \"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8\": container with ID starting with 52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8 not found: ID does not exist" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.444401 4960 scope.go:117] "RemoveContainer" containerID="cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.445699 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d"} err="failed to get container status \"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d\": rpc error: code = NotFound desc = could not find container \"cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d\": container with ID starting with cc570a78bbe08945a7735e34724f29aeb4fd2c831efa0c9d9a11824d81d3168d not found: ID does not exist" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.445739 4960 scope.go:117] "RemoveContainer" containerID="52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.447282 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8"} err="failed to get container status \"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8\": rpc error: code = NotFound desc = could not find container \"52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8\": container with ID starting with 52b2cb420260ea7b18028f29e022c4b0831f516f66728a2aac5ac035dff5a1f8 not found: ID does not exist" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.447955 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.464746 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:12 crc kubenswrapper[4960]: E1002 07:35:12.465486 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-log" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.465512 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-log" Oct 02 07:35:12 crc kubenswrapper[4960]: E1002 07:35:12.465547 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-metadata" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.465554 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-metadata" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.465754 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-metadata" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.465789 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" containerName="nova-metadata-log" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.466967 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.471342 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.472837 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.500134 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.610687 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-config-data\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.610967 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742696d3-3c84-4dd6-a87e-8f898b461b8c-logs\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.611038 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.611110 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvfxk\" (UniqueName: \"kubernetes.io/projected/742696d3-3c84-4dd6-a87e-8f898b461b8c-kube-api-access-zvfxk\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.611130 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.713475 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742696d3-3c84-4dd6-a87e-8f898b461b8c-logs\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.713564 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.713640 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvfxk\" (UniqueName: \"kubernetes.io/projected/742696d3-3c84-4dd6-a87e-8f898b461b8c-kube-api-access-zvfxk\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.713680 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.713740 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-config-data\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.714569 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742696d3-3c84-4dd6-a87e-8f898b461b8c-logs\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.722526 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.722594 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-config-data\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.723640 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.733860 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvfxk\" (UniqueName: \"kubernetes.io/projected/742696d3-3c84-4dd6-a87e-8f898b461b8c-kube-api-access-zvfxk\") pod \"nova-metadata-0\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " pod="openstack/nova-metadata-0" Oct 02 07:35:12 crc kubenswrapper[4960]: I1002 07:35:12.789902 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:13 crc kubenswrapper[4960]: I1002 07:35:13.299622 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:13 crc kubenswrapper[4960]: I1002 07:35:13.395426 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"742696d3-3c84-4dd6-a87e-8f898b461b8c","Type":"ContainerStarted","Data":"76da77235b8ee968ed39cb07e5279bdb07ce7773dac5e8f56ea0537b61adeebb"} Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.279833 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.280504 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7f55241f-c680-4282-9d60-2724a8ba9e7a" containerName="kube-state-metrics" containerID="cri-o://bce0ed98537129efbbd9480f0d9f3cdbd4ddec996d67c43034737ea6ff429c41" gracePeriod=30 Oct 02 07:35:14 crc kubenswrapper[4960]: E1002 07:35:14.358678 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8106fcd_0e3e_4440_bbc7_96cd83d369b5.slice/crio-6389c63ca445a838d188e1446238f1d73f34593d1ba53928214879d4167919c4.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.387487 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f147cb3a-614c-4196-8c40-d1f974dc82a0" path="/var/lib/kubelet/pods/f147cb3a-614c-4196-8c40-d1f974dc82a0/volumes" Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.411783 4960 generic.go:334] "Generic (PLEG): container finished" podID="b8106fcd-0e3e-4440-bbc7-96cd83d369b5" containerID="6389c63ca445a838d188e1446238f1d73f34593d1ba53928214879d4167919c4" exitCode=0 Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.411861 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7nd6x" event={"ID":"b8106fcd-0e3e-4440-bbc7-96cd83d369b5","Type":"ContainerDied","Data":"6389c63ca445a838d188e1446238f1d73f34593d1ba53928214879d4167919c4"} Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.414394 4960 generic.go:334] "Generic (PLEG): container finished" podID="7f55241f-c680-4282-9d60-2724a8ba9e7a" containerID="bce0ed98537129efbbd9480f0d9f3cdbd4ddec996d67c43034737ea6ff429c41" exitCode=2 Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.414448 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7f55241f-c680-4282-9d60-2724a8ba9e7a","Type":"ContainerDied","Data":"bce0ed98537129efbbd9480f0d9f3cdbd4ddec996d67c43034737ea6ff429c41"} Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.423271 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"742696d3-3c84-4dd6-a87e-8f898b461b8c","Type":"ContainerStarted","Data":"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4"} Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.423643 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"742696d3-3c84-4dd6-a87e-8f898b461b8c","Type":"ContainerStarted","Data":"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae"} Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.472847 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.472824504 podStartE2EDuration="2.472824504s" podCreationTimestamp="2025-10-02 07:35:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:14.454852097 +0000 UTC m=+1135.486798384" watchObservedRunningTime="2025-10-02 07:35:14.472824504 +0000 UTC m=+1135.504770781" Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.836792 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.959746 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.959815 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.979478 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plqxz\" (UniqueName: \"kubernetes.io/projected/7f55241f-c680-4282-9d60-2724a8ba9e7a-kube-api-access-plqxz\") pod \"7f55241f-c680-4282-9d60-2724a8ba9e7a\" (UID: \"7f55241f-c680-4282-9d60-2724a8ba9e7a\") " Oct 02 07:35:14 crc kubenswrapper[4960]: I1002 07:35:14.986451 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f55241f-c680-4282-9d60-2724a8ba9e7a-kube-api-access-plqxz" (OuterVolumeSpecName: "kube-api-access-plqxz") pod "7f55241f-c680-4282-9d60-2724a8ba9e7a" (UID: "7f55241f-c680-4282-9d60-2724a8ba9e7a"). InnerVolumeSpecName "kube-api-access-plqxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.082177 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plqxz\" (UniqueName: \"kubernetes.io/projected/7f55241f-c680-4282-9d60-2724a8ba9e7a-kube-api-access-plqxz\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.266024 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.266098 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.315093 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.351220 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.392587 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.440524 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6575d66c-hm94c"] Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.440832 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" podUID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerName="dnsmasq-dns" containerID="cri-o://170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c" gracePeriod=10 Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.477824 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.478558 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7f55241f-c680-4282-9d60-2724a8ba9e7a","Type":"ContainerDied","Data":"a9284cfe1b204fc871c0e83b914dd0e12e54a7730b1450f9b7c3cdd83cbebfd9"} Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.478599 4960 scope.go:117] "RemoveContainer" containerID="bce0ed98537129efbbd9480f0d9f3cdbd4ddec996d67c43034737ea6ff429c41" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.478699 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.480520 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-central-agent" containerID="cri-o://4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6" gracePeriod=30 Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.480684 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-notification-agent" containerID="cri-o://4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a" gracePeriod=30 Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.480704 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="sg-core" containerID="cri-o://208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5" gracePeriod=30 Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.480734 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="proxy-httpd" containerID="cri-o://1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca" gracePeriod=30 Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.526101 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.542184 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.587141 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.619910 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:35:15 crc kubenswrapper[4960]: E1002 07:35:15.620659 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f55241f-c680-4282-9d60-2724a8ba9e7a" containerName="kube-state-metrics" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.620675 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f55241f-c680-4282-9d60-2724a8ba9e7a" containerName="kube-state-metrics" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.620890 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f55241f-c680-4282-9d60-2724a8ba9e7a" containerName="kube-state-metrics" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.621712 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.624602 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.627160 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.644223 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.703572 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.703669 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.703763 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.703834 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvntg\" (UniqueName: \"kubernetes.io/projected/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-api-access-tvntg\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.805343 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvntg\" (UniqueName: \"kubernetes.io/projected/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-api-access-tvntg\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.805893 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.805931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.806004 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.817182 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.818430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.819462 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:15 crc kubenswrapper[4960]: I1002 07:35:15.834936 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvntg\" (UniqueName: \"kubernetes.io/projected/cc53e302-34cd-4f65-9612-bcb47fc6b238-kube-api-access-tvntg\") pod \"kube-state-metrics-0\" (UID: \"cc53e302-34cd-4f65-9612-bcb47fc6b238\") " pod="openstack/kube-state-metrics-0" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.021008 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.043510 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.043754 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.115349 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.116649 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-combined-ca-bundle\") pod \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.116698 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-scripts\") pod \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.116751 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t5p7\" (UniqueName: \"kubernetes.io/projected/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-kube-api-access-5t5p7\") pod \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.116816 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-config-data\") pod \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\" (UID: \"b8106fcd-0e3e-4440-bbc7-96cd83d369b5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.127200 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-kube-api-access-5t5p7" (OuterVolumeSpecName: "kube-api-access-5t5p7") pod "b8106fcd-0e3e-4440-bbc7-96cd83d369b5" (UID: "b8106fcd-0e3e-4440-bbc7-96cd83d369b5"). InnerVolumeSpecName "kube-api-access-5t5p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.127312 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-scripts" (OuterVolumeSpecName: "scripts") pod "b8106fcd-0e3e-4440-bbc7-96cd83d369b5" (UID: "b8106fcd-0e3e-4440-bbc7-96cd83d369b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.148316 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.187164 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-config-data" (OuterVolumeSpecName: "config-data") pod "b8106fcd-0e3e-4440-bbc7-96cd83d369b5" (UID: "b8106fcd-0e3e-4440-bbc7-96cd83d369b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.214265 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8106fcd-0e3e-4440-bbc7-96cd83d369b5" (UID: "b8106fcd-0e3e-4440-bbc7-96cd83d369b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.221380 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.221413 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.221423 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t5p7\" (UniqueName: \"kubernetes.io/projected/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-kube-api-access-5t5p7\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.221433 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8106fcd-0e3e-4440-bbc7-96cd83d369b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.322564 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-sb\") pod \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.322655 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cks5f\" (UniqueName: \"kubernetes.io/projected/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-kube-api-access-cks5f\") pod \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.322821 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-dns-svc\") pod \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.322901 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-config\") pod \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.323119 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-nb\") pod \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\" (UID: \"26d2b40e-4552-4527-b31f-8d5af0b8f7f5\") " Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.336215 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-kube-api-access-cks5f" (OuterVolumeSpecName: "kube-api-access-cks5f") pod "26d2b40e-4552-4527-b31f-8d5af0b8f7f5" (UID: "26d2b40e-4552-4527-b31f-8d5af0b8f7f5"). InnerVolumeSpecName "kube-api-access-cks5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.364987 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f55241f-c680-4282-9d60-2724a8ba9e7a" path="/var/lib/kubelet/pods/7f55241f-c680-4282-9d60-2724a8ba9e7a/volumes" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.425602 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cks5f\" (UniqueName: \"kubernetes.io/projected/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-kube-api-access-cks5f\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.442125 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26d2b40e-4552-4527-b31f-8d5af0b8f7f5" (UID: "26d2b40e-4552-4527-b31f-8d5af0b8f7f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.442320 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-config" (OuterVolumeSpecName: "config") pod "26d2b40e-4552-4527-b31f-8d5af0b8f7f5" (UID: "26d2b40e-4552-4527-b31f-8d5af0b8f7f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.470703 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26d2b40e-4552-4527-b31f-8d5af0b8f7f5" (UID: "26d2b40e-4552-4527-b31f-8d5af0b8f7f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.488172 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26d2b40e-4552-4527-b31f-8d5af0b8f7f5" (UID: "26d2b40e-4552-4527-b31f-8d5af0b8f7f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.493339 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7nd6x" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.493829 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7nd6x" event={"ID":"b8106fcd-0e3e-4440-bbc7-96cd83d369b5","Type":"ContainerDied","Data":"7a94229645b189377e83e8ffd739528a27128b8b300793c4fb16791c23e9cf81"} Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.493896 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a94229645b189377e83e8ffd739528a27128b8b300793c4fb16791c23e9cf81" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.498546 4960 generic.go:334] "Generic (PLEG): container finished" podID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerID="170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c" exitCode=0 Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.498739 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" event={"ID":"26d2b40e-4552-4527-b31f-8d5af0b8f7f5","Type":"ContainerDied","Data":"170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c"} Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.498790 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" event={"ID":"26d2b40e-4552-4527-b31f-8d5af0b8f7f5","Type":"ContainerDied","Data":"792242c7aac159c2c6a9edf7a1ea5e9ea3d2279bf7d3341463ed35b1164f1782"} Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.498819 4960 scope.go:117] "RemoveContainer" containerID="170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.499054 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6575d66c-hm94c" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.503311 4960 generic.go:334] "Generic (PLEG): container finished" podID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerID="1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca" exitCode=0 Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.503350 4960 generic.go:334] "Generic (PLEG): container finished" podID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerID="208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5" exitCode=2 Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.503358 4960 generic.go:334] "Generic (PLEG): container finished" podID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerID="4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6" exitCode=0 Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.504151 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerDied","Data":"1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca"} Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.504239 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerDied","Data":"208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5"} Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.504257 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerDied","Data":"4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6"} Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.528074 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.528942 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.529090 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.529112 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26d2b40e-4552-4527-b31f-8d5af0b8f7f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.540653 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6575d66c-hm94c"] Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.565312 4960 scope.go:117] "RemoveContainer" containerID="8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.583424 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d6575d66c-hm94c"] Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.596245 4960 scope.go:117] "RemoveContainer" containerID="170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c" Oct 02 07:35:16 crc kubenswrapper[4960]: E1002 07:35:16.597064 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c\": container with ID starting with 170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c not found: ID does not exist" containerID="170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.597121 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c"} err="failed to get container status \"170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c\": rpc error: code = NotFound desc = could not find container \"170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c\": container with ID starting with 170881e5ed7e2c5f05283d18790d34a45a7383d802e69b37f0c760036906cd4c not found: ID does not exist" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.597158 4960 scope.go:117] "RemoveContainer" containerID="8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3" Oct 02 07:35:16 crc kubenswrapper[4960]: E1002 07:35:16.597581 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3\": container with ID starting with 8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3 not found: ID does not exist" containerID="8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.597659 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3"} err="failed to get container status \"8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3\": rpc error: code = NotFound desc = could not find container \"8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3\": container with ID starting with 8ed0d0fc7989f78e65d4689fe5a1907609e5f62b867e8d368557728c912878a3 not found: ID does not exist" Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.630952 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.825264 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.825663 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-log" containerID="cri-o://0218a1adbe98c03440bf9db253b26bed017c490bc0e7056af40abc9326f02987" gracePeriod=30 Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.825730 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-api" containerID="cri-o://67c2b2c207b3199ffbe81b5e678f0afca32b8670c354e801d89153f19801fdd5" gracePeriod=30 Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.841040 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.848925 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.849240 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-log" containerID="cri-o://23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae" gracePeriod=30 Oct 02 07:35:16 crc kubenswrapper[4960]: I1002 07:35:16.849508 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-metadata" containerID="cri-o://04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4" gracePeriod=30 Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.515021 4960 generic.go:334] "Generic (PLEG): container finished" podID="37ce726a-c9d7-46af-840f-9df5f9266721" containerID="0218a1adbe98c03440bf9db253b26bed017c490bc0e7056af40abc9326f02987" exitCode=143 Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.515530 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37ce726a-c9d7-46af-840f-9df5f9266721","Type":"ContainerDied","Data":"0218a1adbe98c03440bf9db253b26bed017c490bc0e7056af40abc9326f02987"} Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.520148 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.521880 4960 generic.go:334] "Generic (PLEG): container finished" podID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerID="04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4" exitCode=0 Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.521904 4960 generic.go:334] "Generic (PLEG): container finished" podID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerID="23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae" exitCode=143 Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.521937 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"742696d3-3c84-4dd6-a87e-8f898b461b8c","Type":"ContainerDied","Data":"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4"} Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.521954 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"742696d3-3c84-4dd6-a87e-8f898b461b8c","Type":"ContainerDied","Data":"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae"} Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.521966 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"742696d3-3c84-4dd6-a87e-8f898b461b8c","Type":"ContainerDied","Data":"76da77235b8ee968ed39cb07e5279bdb07ce7773dac5e8f56ea0537b61adeebb"} Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.521998 4960 scope.go:117] "RemoveContainer" containerID="04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.526646 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cc53e302-34cd-4f65-9612-bcb47fc6b238","Type":"ContainerStarted","Data":"6432fbc7969abcf59f5c19be899ee24c40dcee43fa75d13dd04ef618e3b380cf"} Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.526873 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9d6b840e-7ada-4631-b63c-6d1edcf42b21" containerName="nova-scheduler-scheduler" containerID="cri-o://bb013ef65473a5c4ee434838fd66162c8629a9a9ea573902b16eaae58f971dcf" gracePeriod=30 Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.545315 4960 scope.go:117] "RemoveContainer" containerID="23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.580816 4960 scope.go:117] "RemoveContainer" containerID="04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4" Oct 02 07:35:17 crc kubenswrapper[4960]: E1002 07:35:17.581466 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4\": container with ID starting with 04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4 not found: ID does not exist" containerID="04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.581509 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4"} err="failed to get container status \"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4\": rpc error: code = NotFound desc = could not find container \"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4\": container with ID starting with 04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4 not found: ID does not exist" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.581539 4960 scope.go:117] "RemoveContainer" containerID="23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae" Oct 02 07:35:17 crc kubenswrapper[4960]: E1002 07:35:17.583628 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae\": container with ID starting with 23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae not found: ID does not exist" containerID="23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.583678 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae"} err="failed to get container status \"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae\": rpc error: code = NotFound desc = could not find container \"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae\": container with ID starting with 23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae not found: ID does not exist" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.583718 4960 scope.go:117] "RemoveContainer" containerID="04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.584296 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4"} err="failed to get container status \"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4\": rpc error: code = NotFound desc = could not find container \"04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4\": container with ID starting with 04cf20f86eccf0699d819cdf0f6f9af18fca52d6c19e0628d1851e2895d745f4 not found: ID does not exist" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.584311 4960 scope.go:117] "RemoveContainer" containerID="23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.584519 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae"} err="failed to get container status \"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae\": rpc error: code = NotFound desc = could not find container \"23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae\": container with ID starting with 23502d9bb2d9c60ade3dfd08091d10b8cd9aa0f2778725e767008883a25b4cae not found: ID does not exist" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.678871 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvfxk\" (UniqueName: \"kubernetes.io/projected/742696d3-3c84-4dd6-a87e-8f898b461b8c-kube-api-access-zvfxk\") pod \"742696d3-3c84-4dd6-a87e-8f898b461b8c\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.679075 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-combined-ca-bundle\") pod \"742696d3-3c84-4dd6-a87e-8f898b461b8c\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.679147 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-nova-metadata-tls-certs\") pod \"742696d3-3c84-4dd6-a87e-8f898b461b8c\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.679326 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-config-data\") pod \"742696d3-3c84-4dd6-a87e-8f898b461b8c\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.679368 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742696d3-3c84-4dd6-a87e-8f898b461b8c-logs\") pod \"742696d3-3c84-4dd6-a87e-8f898b461b8c\" (UID: \"742696d3-3c84-4dd6-a87e-8f898b461b8c\") " Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.679799 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/742696d3-3c84-4dd6-a87e-8f898b461b8c-logs" (OuterVolumeSpecName: "logs") pod "742696d3-3c84-4dd6-a87e-8f898b461b8c" (UID: "742696d3-3c84-4dd6-a87e-8f898b461b8c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.685946 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742696d3-3c84-4dd6-a87e-8f898b461b8c-kube-api-access-zvfxk" (OuterVolumeSpecName: "kube-api-access-zvfxk") pod "742696d3-3c84-4dd6-a87e-8f898b461b8c" (UID: "742696d3-3c84-4dd6-a87e-8f898b461b8c"). InnerVolumeSpecName "kube-api-access-zvfxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.713530 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-config-data" (OuterVolumeSpecName: "config-data") pod "742696d3-3c84-4dd6-a87e-8f898b461b8c" (UID: "742696d3-3c84-4dd6-a87e-8f898b461b8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.736305 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "742696d3-3c84-4dd6-a87e-8f898b461b8c" (UID: "742696d3-3c84-4dd6-a87e-8f898b461b8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.752180 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "742696d3-3c84-4dd6-a87e-8f898b461b8c" (UID: "742696d3-3c84-4dd6-a87e-8f898b461b8c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.781883 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.781939 4960 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.781958 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742696d3-3c84-4dd6-a87e-8f898b461b8c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.781992 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742696d3-3c84-4dd6-a87e-8f898b461b8c-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:17 crc kubenswrapper[4960]: I1002 07:35:17.782006 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvfxk\" (UniqueName: \"kubernetes.io/projected/742696d3-3c84-4dd6-a87e-8f898b461b8c-kube-api-access-zvfxk\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.345676 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" path="/var/lib/kubelet/pods/26d2b40e-4552-4527-b31f-8d5af0b8f7f5/volumes" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.539766 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.543866 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cc53e302-34cd-4f65-9612-bcb47fc6b238","Type":"ContainerStarted","Data":"0eedf4ceca3e491904a3f5f284ed5e98960e795eb328b9f1bf3fde8950d4c8df"} Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.545227 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.570672 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.579267 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.603450 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:18 crc kubenswrapper[4960]: E1002 07:35:18.604108 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerName="dnsmasq-dns" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604131 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerName="dnsmasq-dns" Oct 02 07:35:18 crc kubenswrapper[4960]: E1002 07:35:18.604155 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-metadata" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604165 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-metadata" Oct 02 07:35:18 crc kubenswrapper[4960]: E1002 07:35:18.604192 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8106fcd-0e3e-4440-bbc7-96cd83d369b5" containerName="nova-manage" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604204 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8106fcd-0e3e-4440-bbc7-96cd83d369b5" containerName="nova-manage" Oct 02 07:35:18 crc kubenswrapper[4960]: E1002 07:35:18.604225 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-log" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604234 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-log" Oct 02 07:35:18 crc kubenswrapper[4960]: E1002 07:35:18.604261 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerName="init" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604273 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerName="init" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604496 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-metadata" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604525 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8106fcd-0e3e-4440-bbc7-96cd83d369b5" containerName="nova-manage" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604543 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="26d2b40e-4552-4527-b31f-8d5af0b8f7f5" containerName="dnsmasq-dns" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.604553 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" containerName="nova-metadata-log" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.605958 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.608597 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.609137 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.611561 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.208670623 podStartE2EDuration="3.611534991s" podCreationTimestamp="2025-10-02 07:35:15 +0000 UTC" firstStartedPulling="2025-10-02 07:35:16.63738237 +0000 UTC m=+1137.669328657" lastFinishedPulling="2025-10-02 07:35:17.040246738 +0000 UTC m=+1138.072193025" observedRunningTime="2025-10-02 07:35:18.591863397 +0000 UTC m=+1139.623809694" watchObservedRunningTime="2025-10-02 07:35:18.611534991 +0000 UTC m=+1139.643481278" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.655117 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.702817 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.703220 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.703370 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvqtz\" (UniqueName: \"kubernetes.io/projected/0d29d929-5ba0-44eb-9604-ef811016b504-kube-api-access-wvqtz\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.703459 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d29d929-5ba0-44eb-9604-ef811016b504-logs\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.703527 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-config-data\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.805133 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvqtz\" (UniqueName: \"kubernetes.io/projected/0d29d929-5ba0-44eb-9604-ef811016b504-kube-api-access-wvqtz\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.805201 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d29d929-5ba0-44eb-9604-ef811016b504-logs\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.805235 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-config-data\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.805288 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.805363 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.807165 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d29d929-5ba0-44eb-9604-ef811016b504-logs\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.813127 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.820410 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.824015 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-config-data\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.828097 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvqtz\" (UniqueName: \"kubernetes.io/projected/0d29d929-5ba0-44eb-9604-ef811016b504-kube-api-access-wvqtz\") pod \"nova-metadata-0\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " pod="openstack/nova-metadata-0" Oct 02 07:35:18 crc kubenswrapper[4960]: I1002 07:35:18.935259 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.435925 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:35:19 crc kubenswrapper[4960]: W1002 07:35:19.448339 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d29d929_5ba0_44eb_9604_ef811016b504.slice/crio-d1e7041b584dbb06cec7e768c7e862df6baa7804787b60ce509841328164c472 WatchSource:0}: Error finding container d1e7041b584dbb06cec7e768c7e862df6baa7804787b60ce509841328164c472: Status 404 returned error can't find the container with id d1e7041b584dbb06cec7e768c7e862df6baa7804787b60ce509841328164c472 Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.556006 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0d29d929-5ba0-44eb-9604-ef811016b504","Type":"ContainerStarted","Data":"d1e7041b584dbb06cec7e768c7e862df6baa7804787b60ce509841328164c472"} Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.558924 4960 generic.go:334] "Generic (PLEG): container finished" podID="334bd1b9-aae9-4744-b800-817bbf01d3d9" containerID="cfcb3e868db40531716a774898dfaab314ca1fc0e1bf24fef47df76f39fcf6d0" exitCode=0 Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.559046 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" event={"ID":"334bd1b9-aae9-4744-b800-817bbf01d3d9","Type":"ContainerDied","Data":"cfcb3e868db40531716a774898dfaab314ca1fc0e1bf24fef47df76f39fcf6d0"} Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.563441 4960 generic.go:334] "Generic (PLEG): container finished" podID="9d6b840e-7ada-4631-b63c-6d1edcf42b21" containerID="bb013ef65473a5c4ee434838fd66162c8629a9a9ea573902b16eaae58f971dcf" exitCode=0 Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.563528 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d6b840e-7ada-4631-b63c-6d1edcf42b21","Type":"ContainerDied","Data":"bb013ef65473a5c4ee434838fd66162c8629a9a9ea573902b16eaae58f971dcf"} Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.696519 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.831962 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqwl6\" (UniqueName: \"kubernetes.io/projected/9d6b840e-7ada-4631-b63c-6d1edcf42b21-kube-api-access-kqwl6\") pod \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.832756 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-combined-ca-bundle\") pod \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.832842 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-config-data\") pod \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\" (UID: \"9d6b840e-7ada-4631-b63c-6d1edcf42b21\") " Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.849444 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d6b840e-7ada-4631-b63c-6d1edcf42b21-kube-api-access-kqwl6" (OuterVolumeSpecName: "kube-api-access-kqwl6") pod "9d6b840e-7ada-4631-b63c-6d1edcf42b21" (UID: "9d6b840e-7ada-4631-b63c-6d1edcf42b21"). InnerVolumeSpecName "kube-api-access-kqwl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.874771 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d6b840e-7ada-4631-b63c-6d1edcf42b21" (UID: "9d6b840e-7ada-4631-b63c-6d1edcf42b21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.884404 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-config-data" (OuterVolumeSpecName: "config-data") pod "9d6b840e-7ada-4631-b63c-6d1edcf42b21" (UID: "9d6b840e-7ada-4631-b63c-6d1edcf42b21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.937456 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqwl6\" (UniqueName: \"kubernetes.io/projected/9d6b840e-7ada-4631-b63c-6d1edcf42b21-kube-api-access-kqwl6\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.937512 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.937527 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b840e-7ada-4631-b63c-6d1edcf42b21-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:19 crc kubenswrapper[4960]: I1002 07:35:19.967149 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.039306 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-config-data\") pod \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.039491 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-combined-ca-bundle\") pod \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.039595 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-log-httpd\") pod \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.039631 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r5df\" (UniqueName: \"kubernetes.io/projected/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-kube-api-access-4r5df\") pod \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.039699 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-scripts\") pod \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.039783 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-run-httpd\") pod \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.039839 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-sg-core-conf-yaml\") pod \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\" (UID: \"c67b5db0-0228-49e1-94e9-fb0cfa0512fd\") " Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.040357 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c67b5db0-0228-49e1-94e9-fb0cfa0512fd" (UID: "c67b5db0-0228-49e1-94e9-fb0cfa0512fd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.040483 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.041177 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c67b5db0-0228-49e1-94e9-fb0cfa0512fd" (UID: "c67b5db0-0228-49e1-94e9-fb0cfa0512fd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.045224 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-scripts" (OuterVolumeSpecName: "scripts") pod "c67b5db0-0228-49e1-94e9-fb0cfa0512fd" (UID: "c67b5db0-0228-49e1-94e9-fb0cfa0512fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.046586 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-kube-api-access-4r5df" (OuterVolumeSpecName: "kube-api-access-4r5df") pod "c67b5db0-0228-49e1-94e9-fb0cfa0512fd" (UID: "c67b5db0-0228-49e1-94e9-fb0cfa0512fd"). InnerVolumeSpecName "kube-api-access-4r5df". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.079098 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c67b5db0-0228-49e1-94e9-fb0cfa0512fd" (UID: "c67b5db0-0228-49e1-94e9-fb0cfa0512fd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.121880 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c67b5db0-0228-49e1-94e9-fb0cfa0512fd" (UID: "c67b5db0-0228-49e1-94e9-fb0cfa0512fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.142815 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.142866 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r5df\" (UniqueName: \"kubernetes.io/projected/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-kube-api-access-4r5df\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.142884 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.142895 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.142907 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.147587 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-config-data" (OuterVolumeSpecName: "config-data") pod "c67b5db0-0228-49e1-94e9-fb0cfa0512fd" (UID: "c67b5db0-0228-49e1-94e9-fb0cfa0512fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.245141 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c67b5db0-0228-49e1-94e9-fb0cfa0512fd-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.344263 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="742696d3-3c84-4dd6-a87e-8f898b461b8c" path="/var/lib/kubelet/pods/742696d3-3c84-4dd6-a87e-8f898b461b8c/volumes" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.584095 4960 generic.go:334] "Generic (PLEG): container finished" podID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerID="4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a" exitCode=0 Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.584253 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerDied","Data":"4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a"} Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.584694 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c67b5db0-0228-49e1-94e9-fb0cfa0512fd","Type":"ContainerDied","Data":"83ae0f48392ca3b18c5d8fe63403a5ff9b355dd70ffc8fe66c41ef801b180e25"} Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.584722 4960 scope.go:117] "RemoveContainer" containerID="1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.584396 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.588632 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9d6b840e-7ada-4631-b63c-6d1edcf42b21","Type":"ContainerDied","Data":"11a97680f7ecd2d7ba8b8f9eb05ad47d236a9a32a905dfae6a0cac893deb697c"} Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.588790 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.601002 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0d29d929-5ba0-44eb-9604-ef811016b504","Type":"ContainerStarted","Data":"3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b"} Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.601084 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0d29d929-5ba0-44eb-9604-ef811016b504","Type":"ContainerStarted","Data":"3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2"} Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.632146 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.657542 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.659659 4960 scope.go:117] "RemoveContainer" containerID="208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.663299 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.663263985 podStartE2EDuration="2.663263985s" podCreationTimestamp="2025-10-02 07:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:20.64975406 +0000 UTC m=+1141.681700347" watchObservedRunningTime="2025-10-02 07:35:20.663263985 +0000 UTC m=+1141.695210272" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.699306 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.699770 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="proxy-httpd" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.699794 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="proxy-httpd" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.699811 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-central-agent" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.699820 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-central-agent" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.699830 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-notification-agent" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.699841 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-notification-agent" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.699868 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d6b840e-7ada-4631-b63c-6d1edcf42b21" containerName="nova-scheduler-scheduler" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.699876 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d6b840e-7ada-4631-b63c-6d1edcf42b21" containerName="nova-scheduler-scheduler" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.699893 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="sg-core" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.699901 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="sg-core" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.700177 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="sg-core" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.700197 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d6b840e-7ada-4631-b63c-6d1edcf42b21" containerName="nova-scheduler-scheduler" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.700212 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="proxy-httpd" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.700227 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-notification-agent" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.700247 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" containerName="ceilometer-central-agent" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.704609 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.720560 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.720835 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.720990 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.727862 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.736556 4960 scope.go:117] "RemoveContainer" containerID="4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.739718 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.752754 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.766190 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.768053 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.771177 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.776511 4960 scope.go:117] "RemoveContainer" containerID="4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.783401 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.818612 4960 scope.go:117] "RemoveContainer" containerID="1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.820482 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca\": container with ID starting with 1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca not found: ID does not exist" containerID="1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.820576 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca"} err="failed to get container status \"1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca\": rpc error: code = NotFound desc = could not find container \"1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca\": container with ID starting with 1fda0f2f29d8f9406aed370dea1a0f2bb84ab4a31bc10b718dcc124628d1a7ca not found: ID does not exist" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.820625 4960 scope.go:117] "RemoveContainer" containerID="208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.821185 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5\": container with ID starting with 208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5 not found: ID does not exist" containerID="208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.821242 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5"} err="failed to get container status \"208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5\": rpc error: code = NotFound desc = could not find container \"208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5\": container with ID starting with 208507c0bca27ab80a716ce870667fdee7e0a2078f8ac3b8666b59dde538b2e5 not found: ID does not exist" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.821286 4960 scope.go:117] "RemoveContainer" containerID="4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.821826 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a\": container with ID starting with 4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a not found: ID does not exist" containerID="4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.821880 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a"} err="failed to get container status \"4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a\": rpc error: code = NotFound desc = could not find container \"4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a\": container with ID starting with 4de1015b526b5526c172f3c71c6a3ab951d0b0d804208c063479776bfe33379a not found: ID does not exist" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.821911 4960 scope.go:117] "RemoveContainer" containerID="4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6" Oct 02 07:35:20 crc kubenswrapper[4960]: E1002 07:35:20.822427 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6\": container with ID starting with 4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6 not found: ID does not exist" containerID="4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.822461 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6"} err="failed to get container status \"4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6\": rpc error: code = NotFound desc = could not find container \"4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6\": container with ID starting with 4729f8cae177502ea303d73cfcaed28f1d52a44087aa351b0bfd9b8288cc34a6 not found: ID does not exist" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.822480 4960 scope.go:117] "RemoveContainer" containerID="bb013ef65473a5c4ee434838fd66162c8629a9a9ea573902b16eaae58f971dcf" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.863227 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-log-httpd\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.863307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz7q5\" (UniqueName: \"kubernetes.io/projected/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-kube-api-access-hz7q5\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.863342 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2b7f\" (UniqueName: \"kubernetes.io/projected/74681e43-80f4-43e9-921c-edb7eec8f795-kube-api-access-z2b7f\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.863369 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-config-data\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.863396 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.863617 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.864875 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-config-data\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.865026 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-scripts\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.865076 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-run-httpd\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.865123 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.865280 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967247 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967318 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967362 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-config-data\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967398 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-scripts\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967420 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-run-httpd\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967443 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967461 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967495 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-log-httpd\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967528 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz7q5\" (UniqueName: \"kubernetes.io/projected/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-kube-api-access-hz7q5\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967564 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2b7f\" (UniqueName: \"kubernetes.io/projected/74681e43-80f4-43e9-921c-edb7eec8f795-kube-api-access-z2b7f\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.967582 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-config-data\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.969595 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-log-httpd\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.969827 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-run-httpd\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.975203 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-scripts\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.976235 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.976626 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.979332 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.979469 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-config-data\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.980751 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.986672 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-config-data\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.991025 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2b7f\" (UniqueName: \"kubernetes.io/projected/74681e43-80f4-43e9-921c-edb7eec8f795-kube-api-access-z2b7f\") pod \"ceilometer-0\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4960]: I1002 07:35:20.991299 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz7q5\" (UniqueName: \"kubernetes.io/projected/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-kube-api-access-hz7q5\") pod \"nova-scheduler-0\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " pod="openstack/nova-scheduler-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.037262 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.064332 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.092796 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.170903 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-scripts\") pod \"334bd1b9-aae9-4744-b800-817bbf01d3d9\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.171064 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc87b\" (UniqueName: \"kubernetes.io/projected/334bd1b9-aae9-4744-b800-817bbf01d3d9-kube-api-access-zc87b\") pod \"334bd1b9-aae9-4744-b800-817bbf01d3d9\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.171148 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-combined-ca-bundle\") pod \"334bd1b9-aae9-4744-b800-817bbf01d3d9\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.171250 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-config-data\") pod \"334bd1b9-aae9-4744-b800-817bbf01d3d9\" (UID: \"334bd1b9-aae9-4744-b800-817bbf01d3d9\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.176878 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-scripts" (OuterVolumeSpecName: "scripts") pod "334bd1b9-aae9-4744-b800-817bbf01d3d9" (UID: "334bd1b9-aae9-4744-b800-817bbf01d3d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.178890 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/334bd1b9-aae9-4744-b800-817bbf01d3d9-kube-api-access-zc87b" (OuterVolumeSpecName: "kube-api-access-zc87b") pod "334bd1b9-aae9-4744-b800-817bbf01d3d9" (UID: "334bd1b9-aae9-4744-b800-817bbf01d3d9"). InnerVolumeSpecName "kube-api-access-zc87b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.202494 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-config-data" (OuterVolumeSpecName: "config-data") pod "334bd1b9-aae9-4744-b800-817bbf01d3d9" (UID: "334bd1b9-aae9-4744-b800-817bbf01d3d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.208277 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "334bd1b9-aae9-4744-b800-817bbf01d3d9" (UID: "334bd1b9-aae9-4744-b800-817bbf01d3d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.274201 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.274247 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc87b\" (UniqueName: \"kubernetes.io/projected/334bd1b9-aae9-4744-b800-817bbf01d3d9-kube-api-access-zc87b\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.274261 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.274271 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/334bd1b9-aae9-4744-b800-817bbf01d3d9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.669241 4960 generic.go:334] "Generic (PLEG): container finished" podID="37ce726a-c9d7-46af-840f-9df5f9266721" containerID="67c2b2c207b3199ffbe81b5e678f0afca32b8670c354e801d89153f19801fdd5" exitCode=0 Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.669329 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37ce726a-c9d7-46af-840f-9df5f9266721","Type":"ContainerDied","Data":"67c2b2c207b3199ffbe81b5e678f0afca32b8670c354e801d89153f19801fdd5"} Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.679848 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.680047 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zxzpz" event={"ID":"334bd1b9-aae9-4744-b800-817bbf01d3d9","Type":"ContainerDied","Data":"d399d7729dcd45522136e836b6c3a052d3cb4fdce84fdd3f83a325f73870cfac"} Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.680108 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d399d7729dcd45522136e836b6c3a052d3cb4fdce84fdd3f83a325f73870cfac" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.692855 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.744927 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.751256 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 07:35:21 crc kubenswrapper[4960]: E1002 07:35:21.751888 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-api" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.751927 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-api" Oct 02 07:35:21 crc kubenswrapper[4960]: E1002 07:35:21.751951 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-log" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.751959 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-log" Oct 02 07:35:21 crc kubenswrapper[4960]: E1002 07:35:21.751968 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="334bd1b9-aae9-4744-b800-817bbf01d3d9" containerName="nova-cell1-conductor-db-sync" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.752000 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="334bd1b9-aae9-4744-b800-817bbf01d3d9" containerName="nova-cell1-conductor-db-sync" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.752334 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-api" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.752364 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" containerName="nova-api-log" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.752378 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="334bd1b9-aae9-4744-b800-817bbf01d3d9" containerName="nova-cell1-conductor-db-sync" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.753541 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.756212 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.765549 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.791635 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891054 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-config-data\") pod \"37ce726a-c9d7-46af-840f-9df5f9266721\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891179 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37ce726a-c9d7-46af-840f-9df5f9266721-logs\") pod \"37ce726a-c9d7-46af-840f-9df5f9266721\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891305 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5n8x\" (UniqueName: \"kubernetes.io/projected/37ce726a-c9d7-46af-840f-9df5f9266721-kube-api-access-t5n8x\") pod \"37ce726a-c9d7-46af-840f-9df5f9266721\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891386 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-combined-ca-bundle\") pod \"37ce726a-c9d7-46af-840f-9df5f9266721\" (UID: \"37ce726a-c9d7-46af-840f-9df5f9266721\") " Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891720 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9qc6\" (UniqueName: \"kubernetes.io/projected/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-kube-api-access-c9qc6\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891757 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37ce726a-c9d7-46af-840f-9df5f9266721-logs" (OuterVolumeSpecName: "logs") pod "37ce726a-c9d7-46af-840f-9df5f9266721" (UID: "37ce726a-c9d7-46af-840f-9df5f9266721"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891807 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.891863 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.892423 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37ce726a-c9d7-46af-840f-9df5f9266721-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.897399 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ce726a-c9d7-46af-840f-9df5f9266721-kube-api-access-t5n8x" (OuterVolumeSpecName: "kube-api-access-t5n8x") pod "37ce726a-c9d7-46af-840f-9df5f9266721" (UID: "37ce726a-c9d7-46af-840f-9df5f9266721"). InnerVolumeSpecName "kube-api-access-t5n8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.923612 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37ce726a-c9d7-46af-840f-9df5f9266721" (UID: "37ce726a-c9d7-46af-840f-9df5f9266721"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.927654 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-config-data" (OuterVolumeSpecName: "config-data") pod "37ce726a-c9d7-46af-840f-9df5f9266721" (UID: "37ce726a-c9d7-46af-840f-9df5f9266721"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.994792 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.994897 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.995020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9qc6\" (UniqueName: \"kubernetes.io/projected/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-kube-api-access-c9qc6\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.995098 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5n8x\" (UniqueName: \"kubernetes.io/projected/37ce726a-c9d7-46af-840f-9df5f9266721-kube-api-access-t5n8x\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.995118 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:21 crc kubenswrapper[4960]: I1002 07:35:21.995130 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ce726a-c9d7-46af-840f-9df5f9266721-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.000090 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.010003 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.018913 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9qc6\" (UniqueName: \"kubernetes.io/projected/c3bd25c3-7b92-4cc8-a683-3ab41bceccf6-kube-api-access-c9qc6\") pod \"nova-cell1-conductor-0\" (UID: \"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.074398 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.340864 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d6b840e-7ada-4631-b63c-6d1edcf42b21" path="/var/lib/kubelet/pods/9d6b840e-7ada-4631-b63c-6d1edcf42b21/volumes" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.342114 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c67b5db0-0228-49e1-94e9-fb0cfa0512fd" path="/var/lib/kubelet/pods/c67b5db0-0228-49e1-94e9-fb0cfa0512fd/volumes" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.580674 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.719077 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerStarted","Data":"4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b"} Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.719646 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerStarted","Data":"dfca8d6d22231cc8db70f60b7e134eb6e16e63cc25b58a781ac431450e8dd6d1"} Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.732882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37ce726a-c9d7-46af-840f-9df5f9266721","Type":"ContainerDied","Data":"182279efd55acf48769e07caeeb2edfcfb583fea19b27fd69983b7a941d3ac10"} Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.732952 4960 scope.go:117] "RemoveContainer" containerID="67c2b2c207b3199ffbe81b5e678f0afca32b8670c354e801d89153f19801fdd5" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.732959 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.742908 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6","Type":"ContainerStarted","Data":"cdd3ea25e9f92ddb7b3d5c7e05da4769f466d68ddf5171fb55116d24613ff04c"} Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.764227 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20","Type":"ContainerStarted","Data":"7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c"} Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.764644 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20","Type":"ContainerStarted","Data":"5c7671d04c1e36f1a9f33b915e62f5efd2c137ec06d4b7f6b3fa7a177f8b5a5b"} Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.768766 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.779569 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.804420 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.807470 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.809934 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.813660 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.813633276 podStartE2EDuration="2.813633276s" podCreationTimestamp="2025-10-02 07:35:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:22.789159853 +0000 UTC m=+1143.821106140" watchObservedRunningTime="2025-10-02 07:35:22.813633276 +0000 UTC m=+1143.845579563" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.816673 4960 scope.go:117] "RemoveContainer" containerID="0218a1adbe98c03440bf9db253b26bed017c490bc0e7056af40abc9326f02987" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.847257 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.914229 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c032dc4-90fe-416d-a79f-79ef391c9444-logs\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.914316 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-config-data\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.914391 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:22 crc kubenswrapper[4960]: I1002 07:35:22.914444 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpjg6\" (UniqueName: \"kubernetes.io/projected/2c032dc4-90fe-416d-a79f-79ef391c9444-kube-api-access-rpjg6\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.016514 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-config-data\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.016611 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.016661 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpjg6\" (UniqueName: \"kubernetes.io/projected/2c032dc4-90fe-416d-a79f-79ef391c9444-kube-api-access-rpjg6\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.016707 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c032dc4-90fe-416d-a79f-79ef391c9444-logs\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.017206 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c032dc4-90fe-416d-a79f-79ef391c9444-logs\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.021461 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.021773 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-config-data\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.038191 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpjg6\" (UniqueName: \"kubernetes.io/projected/2c032dc4-90fe-416d-a79f-79ef391c9444-kube-api-access-rpjg6\") pod \"nova-api-0\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.132244 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.686567 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.781995 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerStarted","Data":"cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25"} Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.787013 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c3bd25c3-7b92-4cc8-a683-3ab41bceccf6","Type":"ContainerStarted","Data":"21239ac2d9462d46dcc6796ad75cf93e97ad076fa09388b26b741da2a5184647"} Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.788279 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.790968 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c032dc4-90fe-416d-a79f-79ef391c9444","Type":"ContainerStarted","Data":"2cd7477ae72a58af9503762dccb1820c51af1606158ecfd019c24fe17ccffa5e"} Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.816063 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.816041437 podStartE2EDuration="2.816041437s" podCreationTimestamp="2025-10-02 07:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:23.814384512 +0000 UTC m=+1144.846330819" watchObservedRunningTime="2025-10-02 07:35:23.816041437 +0000 UTC m=+1144.847987734" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.936397 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:35:23 crc kubenswrapper[4960]: I1002 07:35:23.936955 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:35:24 crc kubenswrapper[4960]: I1002 07:35:24.378923 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ce726a-c9d7-46af-840f-9df5f9266721" path="/var/lib/kubelet/pods/37ce726a-c9d7-46af-840f-9df5f9266721/volumes" Oct 02 07:35:24 crc kubenswrapper[4960]: I1002 07:35:24.806799 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c032dc4-90fe-416d-a79f-79ef391c9444","Type":"ContainerStarted","Data":"c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7"} Oct 02 07:35:24 crc kubenswrapper[4960]: I1002 07:35:24.806871 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c032dc4-90fe-416d-a79f-79ef391c9444","Type":"ContainerStarted","Data":"1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d"} Oct 02 07:35:24 crc kubenswrapper[4960]: I1002 07:35:24.811026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerStarted","Data":"bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8"} Oct 02 07:35:24 crc kubenswrapper[4960]: I1002 07:35:24.848521 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.848490323 podStartE2EDuration="2.848490323s" podCreationTimestamp="2025-10-02 07:35:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:24.838838112 +0000 UTC m=+1145.870784429" watchObservedRunningTime="2025-10-02 07:35:24.848490323 +0000 UTC m=+1145.880436640" Oct 02 07:35:25 crc kubenswrapper[4960]: I1002 07:35:25.832706 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerStarted","Data":"2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b"} Oct 02 07:35:25 crc kubenswrapper[4960]: I1002 07:35:25.833683 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:35:25 crc kubenswrapper[4960]: I1002 07:35:25.878071 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3768158870000002 podStartE2EDuration="5.878042071s" podCreationTimestamp="2025-10-02 07:35:20 +0000 UTC" firstStartedPulling="2025-10-02 07:35:21.712187868 +0000 UTC m=+1142.744134155" lastFinishedPulling="2025-10-02 07:35:25.213414052 +0000 UTC m=+1146.245360339" observedRunningTime="2025-10-02 07:35:25.868941104 +0000 UTC m=+1146.900887391" watchObservedRunningTime="2025-10-02 07:35:25.878042071 +0000 UTC m=+1146.909988358" Oct 02 07:35:26 crc kubenswrapper[4960]: I1002 07:35:26.093904 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 07:35:26 crc kubenswrapper[4960]: I1002 07:35:26.129448 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 07:35:27 crc kubenswrapper[4960]: I1002 07:35:27.123566 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 07:35:28 crc kubenswrapper[4960]: I1002 07:35:28.935759 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:35:28 crc kubenswrapper[4960]: I1002 07:35:28.936062 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:35:29 crc kubenswrapper[4960]: I1002 07:35:29.150163 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:35:29 crc kubenswrapper[4960]: I1002 07:35:29.150257 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:35:29 crc kubenswrapper[4960]: I1002 07:35:29.956994 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:35:29 crc kubenswrapper[4960]: I1002 07:35:29.956987 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:35:31 crc kubenswrapper[4960]: I1002 07:35:31.093611 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 07:35:31 crc kubenswrapper[4960]: I1002 07:35:31.122971 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 07:35:31 crc kubenswrapper[4960]: I1002 07:35:31.960132 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4960]: I1002 07:35:33.133219 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:35:33 crc kubenswrapper[4960]: I1002 07:35:33.133807 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:35:34 crc kubenswrapper[4960]: I1002 07:35:34.217366 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:35:34 crc kubenswrapper[4960]: I1002 07:35:34.217386 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:35:38 crc kubenswrapper[4960]: I1002 07:35:38.951034 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:35:38 crc kubenswrapper[4960]: I1002 07:35:38.955484 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:35:38 crc kubenswrapper[4960]: I1002 07:35:38.960757 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:35:39 crc kubenswrapper[4960]: I1002 07:35:39.001374 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:35:41 crc kubenswrapper[4960]: E1002 07:35:41.690086 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0981f94f_b2f4_46dd_a7c0_f6b761960f30.slice/crio-conmon-c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0981f94f_b2f4_46dd_a7c0_f6b761960f30.slice/crio-c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.814833 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.882442 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-config-data\") pod \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.882702 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-combined-ca-bundle\") pod \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.882778 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5x8f\" (UniqueName: \"kubernetes.io/projected/0981f94f-b2f4-46dd-a7c0-f6b761960f30-kube-api-access-w5x8f\") pod \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\" (UID: \"0981f94f-b2f4-46dd-a7c0-f6b761960f30\") " Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.907144 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0981f94f-b2f4-46dd-a7c0-f6b761960f30-kube-api-access-w5x8f" (OuterVolumeSpecName: "kube-api-access-w5x8f") pod "0981f94f-b2f4-46dd-a7c0-f6b761960f30" (UID: "0981f94f-b2f4-46dd-a7c0-f6b761960f30"). InnerVolumeSpecName "kube-api-access-w5x8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.926627 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0981f94f-b2f4-46dd-a7c0-f6b761960f30" (UID: "0981f94f-b2f4-46dd-a7c0-f6b761960f30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.951719 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-config-data" (OuterVolumeSpecName: "config-data") pod "0981f94f-b2f4-46dd-a7c0-f6b761960f30" (UID: "0981f94f-b2f4-46dd-a7c0-f6b761960f30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.985577 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.985642 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5x8f\" (UniqueName: \"kubernetes.io/projected/0981f94f-b2f4-46dd-a7c0-f6b761960f30-kube-api-access-w5x8f\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:41 crc kubenswrapper[4960]: I1002 07:35:41.985674 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0981f94f-b2f4-46dd-a7c0-f6b761960f30-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.035136 4960 generic.go:334] "Generic (PLEG): container finished" podID="0981f94f-b2f4-46dd-a7c0-f6b761960f30" containerID="c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801" exitCode=137 Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.035239 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.035252 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0981f94f-b2f4-46dd-a7c0-f6b761960f30","Type":"ContainerDied","Data":"c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801"} Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.035366 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0981f94f-b2f4-46dd-a7c0-f6b761960f30","Type":"ContainerDied","Data":"b6e719d5cedd084555a2f721645fe71b21984dba02809ddc708346af5f141f94"} Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.035405 4960 scope.go:117] "RemoveContainer" containerID="c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.068817 4960 scope.go:117] "RemoveContainer" containerID="c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801" Oct 02 07:35:42 crc kubenswrapper[4960]: E1002 07:35:42.069407 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801\": container with ID starting with c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801 not found: ID does not exist" containerID="c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.069465 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801"} err="failed to get container status \"c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801\": rpc error: code = NotFound desc = could not find container \"c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801\": container with ID starting with c3c83b74fe611ba939dbde8f414b343ae97ef9300a183f0db08f90ec816e2801 not found: ID does not exist" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.089605 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.100063 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.131561 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:42 crc kubenswrapper[4960]: E1002 07:35:42.132190 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0981f94f-b2f4-46dd-a7c0-f6b761960f30" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.132223 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0981f94f-b2f4-46dd-a7c0-f6b761960f30" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.132526 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0981f94f-b2f4-46dd-a7c0-f6b761960f30" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.133435 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.137126 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.137666 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.140479 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.143701 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.190336 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq7fk\" (UniqueName: \"kubernetes.io/projected/bc95fd9f-83d9-4731-904a-ba493b31656c-kube-api-access-gq7fk\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.190505 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.190548 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.191156 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.191184 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.292543 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.292931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.292961 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.292996 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.293081 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq7fk\" (UniqueName: \"kubernetes.io/projected/bc95fd9f-83d9-4731-904a-ba493b31656c-kube-api-access-gq7fk\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.298132 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.298155 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.298550 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.299312 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc95fd9f-83d9-4731-904a-ba493b31656c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.313733 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq7fk\" (UniqueName: \"kubernetes.io/projected/bc95fd9f-83d9-4731-904a-ba493b31656c-kube-api-access-gq7fk\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc95fd9f-83d9-4731-904a-ba493b31656c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.345564 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0981f94f-b2f4-46dd-a7c0-f6b761960f30" path="/var/lib/kubelet/pods/0981f94f-b2f4-46dd-a7c0-f6b761960f30/volumes" Oct 02 07:35:42 crc kubenswrapper[4960]: I1002 07:35:42.458945 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:43 crc kubenswrapper[4960]: I1002 07:35:43.004953 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:35:43 crc kubenswrapper[4960]: I1002 07:35:43.055026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bc95fd9f-83d9-4731-904a-ba493b31656c","Type":"ContainerStarted","Data":"6b3167d4c98bfe4912b43f6dbc9e94d47a3385b3bc3a8e62f20e05253dd1a4d2"} Oct 02 07:35:43 crc kubenswrapper[4960]: I1002 07:35:43.136170 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:35:43 crc kubenswrapper[4960]: I1002 07:35:43.136812 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:35:43 crc kubenswrapper[4960]: I1002 07:35:43.141312 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:35:43 crc kubenswrapper[4960]: I1002 07:35:43.143586 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.089393 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bc95fd9f-83d9-4731-904a-ba493b31656c","Type":"ContainerStarted","Data":"a50ec106d4e48fdcaa1cf702d684e10600928fed40af87ad50aeb02b4aa15d7c"} Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.090200 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.095699 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.118803 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.118780207 podStartE2EDuration="2.118780207s" podCreationTimestamp="2025-10-02 07:35:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:44.11304559 +0000 UTC m=+1165.144991887" watchObservedRunningTime="2025-10-02 07:35:44.118780207 +0000 UTC m=+1165.150726494" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.347932 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54bd68d8df-bhlzv"] Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.353378 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.383666 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54bd68d8df-bhlzv"] Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.453527 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n847s\" (UniqueName: \"kubernetes.io/projected/aed2fbca-b0bd-4304-9401-4b4257d33c82-kube-api-access-n847s\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.453588 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-nb\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.453617 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-dns-svc\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.453643 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-config\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.453764 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-sb\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.556993 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-sb\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.557194 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n847s\" (UniqueName: \"kubernetes.io/projected/aed2fbca-b0bd-4304-9401-4b4257d33c82-kube-api-access-n847s\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.557258 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-nb\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.557317 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-dns-svc\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.557384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-config\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.558697 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-dns-svc\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.558798 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-nb\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.559509 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-config\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.559605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-sb\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.586594 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n847s\" (UniqueName: \"kubernetes.io/projected/aed2fbca-b0bd-4304-9401-4b4257d33c82-kube-api-access-n847s\") pod \"dnsmasq-dns-54bd68d8df-bhlzv\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:44 crc kubenswrapper[4960]: I1002 07:35:44.681035 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:45 crc kubenswrapper[4960]: I1002 07:35:45.194729 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54bd68d8df-bhlzv"] Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.111431 4960 generic.go:334] "Generic (PLEG): container finished" podID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerID="0903ea7d79f98e51cafbc9b72c96bbe58ac4986759799d34ccf926f41dd93261" exitCode=0 Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.111548 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" event={"ID":"aed2fbca-b0bd-4304-9401-4b4257d33c82","Type":"ContainerDied","Data":"0903ea7d79f98e51cafbc9b72c96bbe58ac4986759799d34ccf926f41dd93261"} Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.112273 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" event={"ID":"aed2fbca-b0bd-4304-9401-4b4257d33c82","Type":"ContainerStarted","Data":"0e35a0aeccb19160db3b6dc6cc8a314605ca2568ad96b30eedcbc20ad2c3924d"} Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.799188 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.800264 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-central-agent" containerID="cri-o://4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b" gracePeriod=30 Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.801433 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="proxy-httpd" containerID="cri-o://2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b" gracePeriod=30 Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.801496 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-notification-agent" containerID="cri-o://cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25" gracePeriod=30 Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.801682 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="sg-core" containerID="cri-o://bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8" gracePeriod=30 Oct 02 07:35:46 crc kubenswrapper[4960]: I1002 07:35:46.822078 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.186:3000/\": EOF" Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.126042 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" event={"ID":"aed2fbca-b0bd-4304-9401-4b4257d33c82","Type":"ContainerStarted","Data":"bed36969808958469052a09c93af5d4dc9fc02d7e427a6278141ed81476e2882"} Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.126311 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.132308 4960 generic.go:334] "Generic (PLEG): container finished" podID="74681e43-80f4-43e9-921c-edb7eec8f795" containerID="2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b" exitCode=0 Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.132339 4960 generic.go:334] "Generic (PLEG): container finished" podID="74681e43-80f4-43e9-921c-edb7eec8f795" containerID="bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8" exitCode=2 Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.132358 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerDied","Data":"2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b"} Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.132378 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerDied","Data":"bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8"} Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.152313 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" podStartSLOduration=3.152290712 podStartE2EDuration="3.152290712s" podCreationTimestamp="2025-10-02 07:35:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:47.145254082 +0000 UTC m=+1168.177200369" watchObservedRunningTime="2025-10-02 07:35:47.152290712 +0000 UTC m=+1168.184236999" Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.393377 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.393708 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-log" containerID="cri-o://1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d" gracePeriod=30 Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.393913 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-api" containerID="cri-o://c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7" gracePeriod=30 Oct 02 07:35:47 crc kubenswrapper[4960]: I1002 07:35:47.459424 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:48 crc kubenswrapper[4960]: I1002 07:35:48.145277 4960 generic.go:334] "Generic (PLEG): container finished" podID="74681e43-80f4-43e9-921c-edb7eec8f795" containerID="4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b" exitCode=0 Oct 02 07:35:48 crc kubenswrapper[4960]: I1002 07:35:48.145386 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerDied","Data":"4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b"} Oct 02 07:35:48 crc kubenswrapper[4960]: I1002 07:35:48.147660 4960 generic.go:334] "Generic (PLEG): container finished" podID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerID="1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d" exitCode=143 Oct 02 07:35:48 crc kubenswrapper[4960]: I1002 07:35:48.148602 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c032dc4-90fe-416d-a79f-79ef391c9444","Type":"ContainerDied","Data":"1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d"} Oct 02 07:35:49 crc kubenswrapper[4960]: I1002 07:35:49.949174 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.099888 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-run-httpd\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.100060 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-combined-ca-bundle\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.100105 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-scripts\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.100175 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2b7f\" (UniqueName: \"kubernetes.io/projected/74681e43-80f4-43e9-921c-edb7eec8f795-kube-api-access-z2b7f\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.100210 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-log-httpd\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.100308 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-config-data\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.100346 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-sg-core-conf-yaml\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.100433 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-ceilometer-tls-certs\") pod \"74681e43-80f4-43e9-921c-edb7eec8f795\" (UID: \"74681e43-80f4-43e9-921c-edb7eec8f795\") " Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.101200 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.101397 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.108244 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74681e43-80f4-43e9-921c-edb7eec8f795-kube-api-access-z2b7f" (OuterVolumeSpecName: "kube-api-access-z2b7f") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "kube-api-access-z2b7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.114162 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-scripts" (OuterVolumeSpecName: "scripts") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.132117 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.172749 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.172791 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerDied","Data":"cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25"} Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.172888 4960 scope.go:117] "RemoveContainer" containerID="2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.172741 4960 generic.go:334] "Generic (PLEG): container finished" podID="74681e43-80f4-43e9-921c-edb7eec8f795" containerID="cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25" exitCode=0 Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.173117 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74681e43-80f4-43e9-921c-edb7eec8f795","Type":"ContainerDied","Data":"dfca8d6d22231cc8db70f60b7e134eb6e16e63cc25b58a781ac431450e8dd6d1"} Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.182094 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.193779 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.203774 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.203822 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.203841 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.203857 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.203872 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.203886 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2b7f\" (UniqueName: \"kubernetes.io/projected/74681e43-80f4-43e9-921c-edb7eec8f795-kube-api-access-z2b7f\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.203901 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74681e43-80f4-43e9-921c-edb7eec8f795-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.215299 4960 scope.go:117] "RemoveContainer" containerID="bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.229153 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-config-data" (OuterVolumeSpecName: "config-data") pod "74681e43-80f4-43e9-921c-edb7eec8f795" (UID: "74681e43-80f4-43e9-921c-edb7eec8f795"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.238774 4960 scope.go:117] "RemoveContainer" containerID="cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.264688 4960 scope.go:117] "RemoveContainer" containerID="4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.289864 4960 scope.go:117] "RemoveContainer" containerID="2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b" Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.290490 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b\": container with ID starting with 2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b not found: ID does not exist" containerID="2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.290634 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b"} err="failed to get container status \"2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b\": rpc error: code = NotFound desc = could not find container \"2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b\": container with ID starting with 2b678479f0fe19d2e558012fb45203f8faa13f0d444f77643729e7fb0051245b not found: ID does not exist" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.290761 4960 scope.go:117] "RemoveContainer" containerID="bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8" Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.291353 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8\": container with ID starting with bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8 not found: ID does not exist" containerID="bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.291387 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8"} err="failed to get container status \"bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8\": rpc error: code = NotFound desc = could not find container \"bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8\": container with ID starting with bab3d43c6438388828029737829186a89810193ce0e70b2ddc02ef3b8da4d6d8 not found: ID does not exist" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.291412 4960 scope.go:117] "RemoveContainer" containerID="cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25" Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.292137 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25\": container with ID starting with cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25 not found: ID does not exist" containerID="cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.292265 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25"} err="failed to get container status \"cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25\": rpc error: code = NotFound desc = could not find container \"cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25\": container with ID starting with cc9ec3b154aea29440dc2d2ae8dad61f6f6b98bb59d398bb99e1117b86117f25 not found: ID does not exist" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.292359 4960 scope.go:117] "RemoveContainer" containerID="4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b" Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.292810 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b\": container with ID starting with 4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b not found: ID does not exist" containerID="4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.292843 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b"} err="failed to get container status \"4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b\": rpc error: code = NotFound desc = could not find container \"4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b\": container with ID starting with 4fcf80844a34dabc61c9a568dd9d9dcd52c35ed5f3602d410cc5946f076afb0b not found: ID does not exist" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.305936 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74681e43-80f4-43e9-921c-edb7eec8f795-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.534274 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.548943 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.566415 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.567163 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-central-agent" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567188 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-central-agent" Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.567234 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="proxy-httpd" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567243 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="proxy-httpd" Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.567263 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="sg-core" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567271 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="sg-core" Oct 02 07:35:50 crc kubenswrapper[4960]: E1002 07:35:50.567286 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-notification-agent" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567295 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-notification-agent" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567506 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-notification-agent" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567527 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="proxy-httpd" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567549 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="sg-core" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.567574 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" containerName="ceilometer-central-agent" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.570255 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.573598 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.573619 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.574048 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.583366 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.718113 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.718750 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.718785 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5ndw\" (UniqueName: \"kubernetes.io/projected/967ca6d6-7b63-4683-b829-fa7e36d23cc6-kube-api-access-q5ndw\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.718813 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.718894 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-config-data\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.718926 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.718962 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-scripts\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.719413 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-config-data\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822321 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822372 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-scripts\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822533 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822681 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822719 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822757 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5ndw\" (UniqueName: \"kubernetes.io/projected/967ca6d6-7b63-4683-b829-fa7e36d23cc6-kube-api-access-q5ndw\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.822795 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.825651 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.827071 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.831605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.831738 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-scripts\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.832230 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.839558 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-config-data\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.844652 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.851458 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5ndw\" (UniqueName: \"kubernetes.io/projected/967ca6d6-7b63-4683-b829-fa7e36d23cc6-kube-api-access-q5ndw\") pod \"ceilometer-0\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4960]: I1002 07:35:50.986107 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.018580 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.130903 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c032dc4-90fe-416d-a79f-79ef391c9444-logs\") pod \"2c032dc4-90fe-416d-a79f-79ef391c9444\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.131624 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-combined-ca-bundle\") pod \"2c032dc4-90fe-416d-a79f-79ef391c9444\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.131664 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpjg6\" (UniqueName: \"kubernetes.io/projected/2c032dc4-90fe-416d-a79f-79ef391c9444-kube-api-access-rpjg6\") pod \"2c032dc4-90fe-416d-a79f-79ef391c9444\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.131722 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-config-data\") pod \"2c032dc4-90fe-416d-a79f-79ef391c9444\" (UID: \"2c032dc4-90fe-416d-a79f-79ef391c9444\") " Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.131909 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c032dc4-90fe-416d-a79f-79ef391c9444-logs" (OuterVolumeSpecName: "logs") pod "2c032dc4-90fe-416d-a79f-79ef391c9444" (UID: "2c032dc4-90fe-416d-a79f-79ef391c9444"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.132360 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c032dc4-90fe-416d-a79f-79ef391c9444-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.139811 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c032dc4-90fe-416d-a79f-79ef391c9444-kube-api-access-rpjg6" (OuterVolumeSpecName: "kube-api-access-rpjg6") pod "2c032dc4-90fe-416d-a79f-79ef391c9444" (UID: "2c032dc4-90fe-416d-a79f-79ef391c9444"). InnerVolumeSpecName "kube-api-access-rpjg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.171636 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c032dc4-90fe-416d-a79f-79ef391c9444" (UID: "2c032dc4-90fe-416d-a79f-79ef391c9444"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.193307 4960 generic.go:334] "Generic (PLEG): container finished" podID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerID="c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7" exitCode=0 Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.193368 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c032dc4-90fe-416d-a79f-79ef391c9444","Type":"ContainerDied","Data":"c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7"} Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.193406 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c032dc4-90fe-416d-a79f-79ef391c9444","Type":"ContainerDied","Data":"2cd7477ae72a58af9503762dccb1820c51af1606158ecfd019c24fe17ccffa5e"} Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.193432 4960 scope.go:117] "RemoveContainer" containerID="c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.193634 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.197190 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-config-data" (OuterVolumeSpecName: "config-data") pod "2c032dc4-90fe-416d-a79f-79ef391c9444" (UID: "2c032dc4-90fe-416d-a79f-79ef391c9444"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.230448 4960 scope.go:117] "RemoveContainer" containerID="1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.235414 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.235470 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpjg6\" (UniqueName: \"kubernetes.io/projected/2c032dc4-90fe-416d-a79f-79ef391c9444-kube-api-access-rpjg6\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.235486 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c032dc4-90fe-416d-a79f-79ef391c9444-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.259548 4960 scope.go:117] "RemoveContainer" containerID="c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7" Oct 02 07:35:51 crc kubenswrapper[4960]: E1002 07:35:51.260409 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7\": container with ID starting with c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7 not found: ID does not exist" containerID="c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.260460 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7"} err="failed to get container status \"c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7\": rpc error: code = NotFound desc = could not find container \"c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7\": container with ID starting with c9349982b5257709a11ce291f1fde05037fa5359145e5362b4768fe8b95177c7 not found: ID does not exist" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.260499 4960 scope.go:117] "RemoveContainer" containerID="1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d" Oct 02 07:35:51 crc kubenswrapper[4960]: E1002 07:35:51.261331 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d\": container with ID starting with 1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d not found: ID does not exist" containerID="1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.261401 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d"} err="failed to get container status \"1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d\": rpc error: code = NotFound desc = could not find container \"1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d\": container with ID starting with 1f42c0ae1c76c49d763495d31aa0a7ac12cf4f9d70edc6582508f34ba37e469d not found: ID does not exist" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.536893 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:51 crc kubenswrapper[4960]: W1002 07:35:51.547434 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod967ca6d6_7b63_4683_b829_fa7e36d23cc6.slice/crio-b8428fd482a7b0fc9dad3a24bc444ed126410c796a87ad37c37e11b739c6e837 WatchSource:0}: Error finding container b8428fd482a7b0fc9dad3a24bc444ed126410c796a87ad37c37e11b739c6e837: Status 404 returned error can't find the container with id b8428fd482a7b0fc9dad3a24bc444ed126410c796a87ad37c37e11b739c6e837 Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.560478 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.563127 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.583502 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:51 crc kubenswrapper[4960]: E1002 07:35:51.584699 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-api" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.584721 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-api" Oct 02 07:35:51 crc kubenswrapper[4960]: E1002 07:35:51.584932 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-log" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.584938 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-log" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.585213 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-log" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.585235 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" containerName="nova-api-api" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.587733 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.592867 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.593198 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.593511 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.603469 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.644284 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx8gx\" (UniqueName: \"kubernetes.io/projected/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-kube-api-access-wx8gx\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.644366 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-config-data\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.644590 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-public-tls-certs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.644643 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-logs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.644838 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.645112 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.747008 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx8gx\" (UniqueName: \"kubernetes.io/projected/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-kube-api-access-wx8gx\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.747103 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-config-data\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.747139 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-public-tls-certs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.747189 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-logs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.747235 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.747276 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.748962 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-logs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.752326 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.752529 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-config-data\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.754186 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.759344 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-public-tls-certs\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.771663 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx8gx\" (UniqueName: \"kubernetes.io/projected/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-kube-api-access-wx8gx\") pod \"nova-api-0\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " pod="openstack/nova-api-0" Oct 02 07:35:51 crc kubenswrapper[4960]: I1002 07:35:51.920788 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:35:52 crc kubenswrapper[4960]: I1002 07:35:52.206241 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerStarted","Data":"b8428fd482a7b0fc9dad3a24bc444ed126410c796a87ad37c37e11b739c6e837"} Oct 02 07:35:52 crc kubenswrapper[4960]: I1002 07:35:52.343250 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c032dc4-90fe-416d-a79f-79ef391c9444" path="/var/lib/kubelet/pods/2c032dc4-90fe-416d-a79f-79ef391c9444/volumes" Oct 02 07:35:52 crc kubenswrapper[4960]: I1002 07:35:52.345034 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74681e43-80f4-43e9-921c-edb7eec8f795" path="/var/lib/kubelet/pods/74681e43-80f4-43e9-921c-edb7eec8f795/volumes" Oct 02 07:35:52 crc kubenswrapper[4960]: I1002 07:35:52.389432 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:35:52 crc kubenswrapper[4960]: I1002 07:35:52.459790 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:52 crc kubenswrapper[4960]: I1002 07:35:52.494920 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.231289 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a968da5e-fbe3-4c48-a71f-c1b17d8731d8","Type":"ContainerStarted","Data":"f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3"} Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.231854 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a968da5e-fbe3-4c48-a71f-c1b17d8731d8","Type":"ContainerStarted","Data":"c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913"} Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.231909 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a968da5e-fbe3-4c48-a71f-c1b17d8731d8","Type":"ContainerStarted","Data":"af1acfe163a05d6032619e78d8a3c3f009f0f596d7f34a882cd6ea2e569a62c4"} Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.235749 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerStarted","Data":"4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d"} Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.235843 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerStarted","Data":"4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1"} Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.267004 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.274745 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.274714478 podStartE2EDuration="2.274714478s" podCreationTimestamp="2025-10-02 07:35:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:53.263482842 +0000 UTC m=+1174.295429119" watchObservedRunningTime="2025-10-02 07:35:53.274714478 +0000 UTC m=+1174.306660765" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.524681 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-k5544"] Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.526930 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.531409 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.532053 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.536165 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-k5544"] Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.585869 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-config-data\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.585948 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxv4f\" (UniqueName: \"kubernetes.io/projected/dff90544-5bf1-4efd-becf-7f5f271f5de6-kube-api-access-dxv4f\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.586030 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.586134 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-scripts\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.687726 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-scripts\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.687819 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-config-data\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.687891 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxv4f\" (UniqueName: \"kubernetes.io/projected/dff90544-5bf1-4efd-becf-7f5f271f5de6-kube-api-access-dxv4f\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.687919 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.693801 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-scripts\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.694667 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-config-data\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.697831 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.714422 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxv4f\" (UniqueName: \"kubernetes.io/projected/dff90544-5bf1-4efd-becf-7f5f271f5de6-kube-api-access-dxv4f\") pod \"nova-cell1-cell-mapping-k5544\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:53 crc kubenswrapper[4960]: I1002 07:35:53.852487 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:35:54 crc kubenswrapper[4960]: I1002 07:35:54.250736 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerStarted","Data":"450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b"} Oct 02 07:35:54 crc kubenswrapper[4960]: I1002 07:35:54.398244 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-k5544"] Oct 02 07:35:54 crc kubenswrapper[4960]: W1002 07:35:54.407929 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddff90544_5bf1_4efd_becf_7f5f271f5de6.slice/crio-07c3e66abdd9bbefc7d087313e76e0cdc67d777fa13fdc592744dbc74bc34e96 WatchSource:0}: Error finding container 07c3e66abdd9bbefc7d087313e76e0cdc67d777fa13fdc592744dbc74bc34e96: Status 404 returned error can't find the container with id 07c3e66abdd9bbefc7d087313e76e0cdc67d777fa13fdc592744dbc74bc34e96 Oct 02 07:35:54 crc kubenswrapper[4960]: I1002 07:35:54.683144 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:35:54 crc kubenswrapper[4960]: I1002 07:35:54.757201 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69775dfb85-5zwxs"] Oct 02 07:35:54 crc kubenswrapper[4960]: I1002 07:35:54.757578 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" podUID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerName="dnsmasq-dns" containerID="cri-o://b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a" gracePeriod=10 Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.270451 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.283554 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k5544" event={"ID":"dff90544-5bf1-4efd-becf-7f5f271f5de6","Type":"ContainerStarted","Data":"315df6ccfeccbdb10e579ff9ab235bfeb9cb06d2a0b0bc10bae78531979b53e8"} Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.284119 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k5544" event={"ID":"dff90544-5bf1-4efd-becf-7f5f271f5de6","Type":"ContainerStarted","Data":"07c3e66abdd9bbefc7d087313e76e0cdc67d777fa13fdc592744dbc74bc34e96"} Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.294695 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerStarted","Data":"9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e"} Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.294911 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.298795 4960 generic.go:334] "Generic (PLEG): container finished" podID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerID="b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a" exitCode=0 Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.298858 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" event={"ID":"fc70035e-3a13-476f-a5de-e13255cd4a5a","Type":"ContainerDied","Data":"b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a"} Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.298897 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" event={"ID":"fc70035e-3a13-476f-a5de-e13255cd4a5a","Type":"ContainerDied","Data":"28671a7eca61b82f9cf5f42cd3bde6c2a3a7e1ac7046c0bf35fb67afd84d8adc"} Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.298917 4960 scope.go:117] "RemoveContainer" containerID="b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.298864 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69775dfb85-5zwxs" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.328893 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-sb\") pod \"fc70035e-3a13-476f-a5de-e13255cd4a5a\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.328960 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-config\") pod \"fc70035e-3a13-476f-a5de-e13255cd4a5a\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.329487 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-nb\") pod \"fc70035e-3a13-476f-a5de-e13255cd4a5a\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.329542 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-dns-svc\") pod \"fc70035e-3a13-476f-a5de-e13255cd4a5a\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.329613 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48sjw\" (UniqueName: \"kubernetes.io/projected/fc70035e-3a13-476f-a5de-e13255cd4a5a-kube-api-access-48sjw\") pod \"fc70035e-3a13-476f-a5de-e13255cd4a5a\" (UID: \"fc70035e-3a13-476f-a5de-e13255cd4a5a\") " Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.329934 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-k5544" podStartSLOduration=2.329901694 podStartE2EDuration="2.329901694s" podCreationTimestamp="2025-10-02 07:35:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:55.308780443 +0000 UTC m=+1176.340726720" watchObservedRunningTime="2025-10-02 07:35:55.329901694 +0000 UTC m=+1176.361847981" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.335955 4960 scope.go:117] "RemoveContainer" containerID="2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.350293 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.082980687 podStartE2EDuration="5.350262877s" podCreationTimestamp="2025-10-02 07:35:50 +0000 UTC" firstStartedPulling="2025-10-02 07:35:51.560439182 +0000 UTC m=+1172.592385469" lastFinishedPulling="2025-10-02 07:35:54.827721362 +0000 UTC m=+1175.859667659" observedRunningTime="2025-10-02 07:35:55.340640288 +0000 UTC m=+1176.372586575" watchObservedRunningTime="2025-10-02 07:35:55.350262877 +0000 UTC m=+1176.382209164" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.352419 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc70035e-3a13-476f-a5de-e13255cd4a5a-kube-api-access-48sjw" (OuterVolumeSpecName: "kube-api-access-48sjw") pod "fc70035e-3a13-476f-a5de-e13255cd4a5a" (UID: "fc70035e-3a13-476f-a5de-e13255cd4a5a"). InnerVolumeSpecName "kube-api-access-48sjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.408857 4960 scope.go:117] "RemoveContainer" containerID="b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.409397 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc70035e-3a13-476f-a5de-e13255cd4a5a" (UID: "fc70035e-3a13-476f-a5de-e13255cd4a5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4960]: E1002 07:35:55.412512 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a\": container with ID starting with b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a not found: ID does not exist" containerID="b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.412570 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a"} err="failed to get container status \"b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a\": rpc error: code = NotFound desc = could not find container \"b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a\": container with ID starting with b13ce176694aeaea453c4b1fba51fa1a629bd5687e03f7ab065ea2681e2c9c0a not found: ID does not exist" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.412643 4960 scope.go:117] "RemoveContainer" containerID="2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89" Oct 02 07:35:55 crc kubenswrapper[4960]: E1002 07:35:55.413352 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89\": container with ID starting with 2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89 not found: ID does not exist" containerID="2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.413432 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89"} err="failed to get container status \"2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89\": rpc error: code = NotFound desc = could not find container \"2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89\": container with ID starting with 2c9b541321b2af643a6c8ba675fd54c1990ac96300d23ac5f296184b1b969c89 not found: ID does not exist" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.425454 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fc70035e-3a13-476f-a5de-e13255cd4a5a" (UID: "fc70035e-3a13-476f-a5de-e13255cd4a5a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.432333 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48sjw\" (UniqueName: \"kubernetes.io/projected/fc70035e-3a13-476f-a5de-e13255cd4a5a-kube-api-access-48sjw\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.432359 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.432369 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.440054 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fc70035e-3a13-476f-a5de-e13255cd4a5a" (UID: "fc70035e-3a13-476f-a5de-e13255cd4a5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.440052 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-config" (OuterVolumeSpecName: "config") pod "fc70035e-3a13-476f-a5de-e13255cd4a5a" (UID: "fc70035e-3a13-476f-a5de-e13255cd4a5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.534391 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.534449 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc70035e-3a13-476f-a5de-e13255cd4a5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.650706 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69775dfb85-5zwxs"] Oct 02 07:35:55 crc kubenswrapper[4960]: I1002 07:35:55.661363 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69775dfb85-5zwxs"] Oct 02 07:35:56 crc kubenswrapper[4960]: I1002 07:35:56.348791 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc70035e-3a13-476f-a5de-e13255cd4a5a" path="/var/lib/kubelet/pods/fc70035e-3a13-476f-a5de-e13255cd4a5a/volumes" Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.150087 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.151068 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.151144 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.152234 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae97637d73736b75a47ea930cb3cf4b53e5a377c7969711a5fc9d3103045505c"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.152301 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://ae97637d73736b75a47ea930cb3cf4b53e5a377c7969711a5fc9d3103045505c" gracePeriod=600 Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.351273 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="ae97637d73736b75a47ea930cb3cf4b53e5a377c7969711a5fc9d3103045505c" exitCode=0 Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.351437 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"ae97637d73736b75a47ea930cb3cf4b53e5a377c7969711a5fc9d3103045505c"} Oct 02 07:35:59 crc kubenswrapper[4960]: I1002 07:35:59.351953 4960 scope.go:117] "RemoveContainer" containerID="5fc62ae8f7b8e010f4106ce66a46ce004d1ca089113eca9aa199c28f45628180" Oct 02 07:36:00 crc kubenswrapper[4960]: I1002 07:36:00.375990 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"f5bb746b73018c3c6e2b85b664f4b1896310fbe4f05d3269e3eee3a9625ffe9d"} Oct 02 07:36:00 crc kubenswrapper[4960]: I1002 07:36:00.382925 4960 generic.go:334] "Generic (PLEG): container finished" podID="dff90544-5bf1-4efd-becf-7f5f271f5de6" containerID="315df6ccfeccbdb10e579ff9ab235bfeb9cb06d2a0b0bc10bae78531979b53e8" exitCode=0 Oct 02 07:36:00 crc kubenswrapper[4960]: I1002 07:36:00.383026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k5544" event={"ID":"dff90544-5bf1-4efd-becf-7f5f271f5de6","Type":"ContainerDied","Data":"315df6ccfeccbdb10e579ff9ab235bfeb9cb06d2a0b0bc10bae78531979b53e8"} Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.812665 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.898141 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-config-data\") pod \"dff90544-5bf1-4efd-becf-7f5f271f5de6\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.898455 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-scripts\") pod \"dff90544-5bf1-4efd-becf-7f5f271f5de6\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.898596 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxv4f\" (UniqueName: \"kubernetes.io/projected/dff90544-5bf1-4efd-becf-7f5f271f5de6-kube-api-access-dxv4f\") pod \"dff90544-5bf1-4efd-becf-7f5f271f5de6\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.898649 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-combined-ca-bundle\") pod \"dff90544-5bf1-4efd-becf-7f5f271f5de6\" (UID: \"dff90544-5bf1-4efd-becf-7f5f271f5de6\") " Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.906490 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-scripts" (OuterVolumeSpecName: "scripts") pod "dff90544-5bf1-4efd-becf-7f5f271f5de6" (UID: "dff90544-5bf1-4efd-becf-7f5f271f5de6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.907524 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dff90544-5bf1-4efd-becf-7f5f271f5de6-kube-api-access-dxv4f" (OuterVolumeSpecName: "kube-api-access-dxv4f") pod "dff90544-5bf1-4efd-becf-7f5f271f5de6" (UID: "dff90544-5bf1-4efd-becf-7f5f271f5de6"). InnerVolumeSpecName "kube-api-access-dxv4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.921469 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.922591 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.936712 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-config-data" (OuterVolumeSpecName: "config-data") pod "dff90544-5bf1-4efd-becf-7f5f271f5de6" (UID: "dff90544-5bf1-4efd-becf-7f5f271f5de6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:01 crc kubenswrapper[4960]: I1002 07:36:01.941030 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dff90544-5bf1-4efd-becf-7f5f271f5de6" (UID: "dff90544-5bf1-4efd-becf-7f5f271f5de6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.001746 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.002309 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.002372 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxv4f\" (UniqueName: \"kubernetes.io/projected/dff90544-5bf1-4efd-becf-7f5f271f5de6-kube-api-access-dxv4f\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.002411 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dff90544-5bf1-4efd-becf-7f5f271f5de6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.405187 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k5544" event={"ID":"dff90544-5bf1-4efd-becf-7f5f271f5de6","Type":"ContainerDied","Data":"07c3e66abdd9bbefc7d087313e76e0cdc67d777fa13fdc592744dbc74bc34e96"} Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.405254 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07c3e66abdd9bbefc7d087313e76e0cdc67d777fa13fdc592744dbc74bc34e96" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.405297 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k5544" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.628160 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.650305 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.650706 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" containerName="nova-scheduler-scheduler" containerID="cri-o://7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c" gracePeriod=30 Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.671361 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.671719 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-log" containerID="cri-o://3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2" gracePeriod=30 Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.671905 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-metadata" containerID="cri-o://3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b" gracePeriod=30 Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.939220 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:02 crc kubenswrapper[4960]: I1002 07:36:02.939290 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:03 crc kubenswrapper[4960]: I1002 07:36:03.419602 4960 generic.go:334] "Generic (PLEG): container finished" podID="0d29d929-5ba0-44eb-9604-ef811016b504" containerID="3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2" exitCode=143 Oct 02 07:36:03 crc kubenswrapper[4960]: I1002 07:36:03.419708 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0d29d929-5ba0-44eb-9604-ef811016b504","Type":"ContainerDied","Data":"3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2"} Oct 02 07:36:03 crc kubenswrapper[4960]: I1002 07:36:03.420015 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-api" containerID="cri-o://f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3" gracePeriod=30 Oct 02 07:36:03 crc kubenswrapper[4960]: I1002 07:36:03.419968 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-log" containerID="cri-o://c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913" gracePeriod=30 Oct 02 07:36:04 crc kubenswrapper[4960]: I1002 07:36:04.436430 4960 generic.go:334] "Generic (PLEG): container finished" podID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerID="c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913" exitCode=143 Oct 02 07:36:04 crc kubenswrapper[4960]: I1002 07:36:04.436580 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a968da5e-fbe3-4c48-a71f-c1b17d8731d8","Type":"ContainerDied","Data":"c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913"} Oct 02 07:36:05 crc kubenswrapper[4960]: I1002 07:36:05.839060 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": read tcp 10.217.0.2:59630->10.217.0.185:8775: read: connection reset by peer" Oct 02 07:36:05 crc kubenswrapper[4960]: I1002 07:36:05.839060 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": read tcp 10.217.0.2:59640->10.217.0.185:8775: read: connection reset by peer" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.083722 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.204062 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-config-data\") pod \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.204350 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-combined-ca-bundle\") pod \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.204390 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz7q5\" (UniqueName: \"kubernetes.io/projected/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-kube-api-access-hz7q5\") pod \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\" (UID: \"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.213333 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-kube-api-access-hz7q5" (OuterVolumeSpecName: "kube-api-access-hz7q5") pod "f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" (UID: "f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20"). InnerVolumeSpecName "kube-api-access-hz7q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.284285 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" (UID: "f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.290099 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-config-data" (OuterVolumeSpecName: "config-data") pod "f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" (UID: "f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.310168 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.310432 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz7q5\" (UniqueName: \"kubernetes.io/projected/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-kube-api-access-hz7q5\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.310537 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.321587 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.411864 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-config-data\") pod \"0d29d929-5ba0-44eb-9604-ef811016b504\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.412047 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvqtz\" (UniqueName: \"kubernetes.io/projected/0d29d929-5ba0-44eb-9604-ef811016b504-kube-api-access-wvqtz\") pod \"0d29d929-5ba0-44eb-9604-ef811016b504\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.412192 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-nova-metadata-tls-certs\") pod \"0d29d929-5ba0-44eb-9604-ef811016b504\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.412240 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d29d929-5ba0-44eb-9604-ef811016b504-logs\") pod \"0d29d929-5ba0-44eb-9604-ef811016b504\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.412329 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-combined-ca-bundle\") pod \"0d29d929-5ba0-44eb-9604-ef811016b504\" (UID: \"0d29d929-5ba0-44eb-9604-ef811016b504\") " Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.417182 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d29d929-5ba0-44eb-9604-ef811016b504-logs" (OuterVolumeSpecName: "logs") pod "0d29d929-5ba0-44eb-9604-ef811016b504" (UID: "0d29d929-5ba0-44eb-9604-ef811016b504"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.457195 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d29d929-5ba0-44eb-9604-ef811016b504-kube-api-access-wvqtz" (OuterVolumeSpecName: "kube-api-access-wvqtz") pod "0d29d929-5ba0-44eb-9604-ef811016b504" (UID: "0d29d929-5ba0-44eb-9604-ef811016b504"). InnerVolumeSpecName "kube-api-access-wvqtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.479825 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d29d929-5ba0-44eb-9604-ef811016b504" (UID: "0d29d929-5ba0-44eb-9604-ef811016b504"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.491319 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-config-data" (OuterVolumeSpecName: "config-data") pod "0d29d929-5ba0-44eb-9604-ef811016b504" (UID: "0d29d929-5ba0-44eb-9604-ef811016b504"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.509719 4960 generic.go:334] "Generic (PLEG): container finished" podID="0d29d929-5ba0-44eb-9604-ef811016b504" containerID="3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b" exitCode=0 Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.510222 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.511198 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0d29d929-5ba0-44eb-9604-ef811016b504","Type":"ContainerDied","Data":"3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b"} Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.511278 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0d29d929-5ba0-44eb-9604-ef811016b504","Type":"ContainerDied","Data":"d1e7041b584dbb06cec7e768c7e862df6baa7804787b60ce509841328164c472"} Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.511298 4960 scope.go:117] "RemoveContainer" containerID="3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.514419 4960 generic.go:334] "Generic (PLEG): container finished" podID="f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" containerID="7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c" exitCode=0 Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.514459 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20","Type":"ContainerDied","Data":"7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c"} Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.514488 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20","Type":"ContainerDied","Data":"5c7671d04c1e36f1a9f33b915e62f5efd2c137ec06d4b7f6b3fa7a177f8b5a5b"} Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.514914 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.514942 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.514954 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvqtz\" (UniqueName: \"kubernetes.io/projected/0d29d929-5ba0-44eb-9604-ef811016b504-kube-api-access-wvqtz\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.514966 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d29d929-5ba0-44eb-9604-ef811016b504-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.515057 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.529752 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0d29d929-5ba0-44eb-9604-ef811016b504" (UID: "0d29d929-5ba0-44eb-9604-ef811016b504"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.548819 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.552680 4960 scope.go:117] "RemoveContainer" containerID="3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.563663 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.582635 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.583171 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff90544-5bf1-4efd-becf-7f5f271f5de6" containerName="nova-manage" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583189 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff90544-5bf1-4efd-becf-7f5f271f5de6" containerName="nova-manage" Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.583212 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerName="dnsmasq-dns" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583218 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerName="dnsmasq-dns" Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.583234 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-log" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583240 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-log" Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.583259 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerName="init" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583265 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerName="init" Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.583284 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-metadata" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583290 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-metadata" Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.583306 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" containerName="nova-scheduler-scheduler" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583311 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" containerName="nova-scheduler-scheduler" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583490 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-metadata" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583503 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" containerName="nova-scheduler-scheduler" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583523 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="dff90544-5bf1-4efd-becf-7f5f271f5de6" containerName="nova-manage" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583530 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc70035e-3a13-476f-a5de-e13255cd4a5a" containerName="dnsmasq-dns" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.583537 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" containerName="nova-metadata-log" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.584300 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.590281 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.592433 4960 scope.go:117] "RemoveContainer" containerID="3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.602196 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.606573 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b\": container with ID starting with 3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b not found: ID does not exist" containerID="3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.606807 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b"} err="failed to get container status \"3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b\": rpc error: code = NotFound desc = could not find container \"3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b\": container with ID starting with 3d06fd640b857c414e211b3c30be1bf33aea76f69847785d57846b2c7709687b not found: ID does not exist" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.606904 4960 scope.go:117] "RemoveContainer" containerID="3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2" Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.608255 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2\": container with ID starting with 3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2 not found: ID does not exist" containerID="3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.608311 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2"} err="failed to get container status \"3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2\": rpc error: code = NotFound desc = could not find container \"3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2\": container with ID starting with 3376ddead750dd17cc0ba3c16f63e15f8c58410ab5a5bd82e652235f8d1627f2 not found: ID does not exist" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.608344 4960 scope.go:117] "RemoveContainer" containerID="7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.617284 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189dd145-b9b7-4dc5-9fae-a44b359f8748-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.617359 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189dd145-b9b7-4dc5-9fae-a44b359f8748-config-data\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.617426 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r65v\" (UniqueName: \"kubernetes.io/projected/189dd145-b9b7-4dc5-9fae-a44b359f8748-kube-api-access-4r65v\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.617586 4960 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d29d929-5ba0-44eb-9604-ef811016b504-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.674343 4960 scope.go:117] "RemoveContainer" containerID="7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c" Oct 02 07:36:06 crc kubenswrapper[4960]: E1002 07:36:06.675328 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c\": container with ID starting with 7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c not found: ID does not exist" containerID="7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.675404 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c"} err="failed to get container status \"7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c\": rpc error: code = NotFound desc = could not find container \"7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c\": container with ID starting with 7f7c3cff9fe6fc33d985078dbabf5cdaf3415cd450d57e8abbc141c1df1d413c not found: ID does not exist" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.720129 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r65v\" (UniqueName: \"kubernetes.io/projected/189dd145-b9b7-4dc5-9fae-a44b359f8748-kube-api-access-4r65v\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.720969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189dd145-b9b7-4dc5-9fae-a44b359f8748-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.721845 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189dd145-b9b7-4dc5-9fae-a44b359f8748-config-data\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.725290 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189dd145-b9b7-4dc5-9fae-a44b359f8748-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.738476 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/189dd145-b9b7-4dc5-9fae-a44b359f8748-config-data\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.739275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r65v\" (UniqueName: \"kubernetes.io/projected/189dd145-b9b7-4dc5-9fae-a44b359f8748-kube-api-access-4r65v\") pod \"nova-scheduler-0\" (UID: \"189dd145-b9b7-4dc5-9fae-a44b359f8748\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.868037 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.877917 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.888071 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.890243 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.893011 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.893285 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.894685 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.928672 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.928740 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a94097fa-b407-419c-9fad-5078a1f8967f-logs\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.928857 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-config-data\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.928887 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcdzv\" (UniqueName: \"kubernetes.io/projected/a94097fa-b407-419c-9fad-5078a1f8967f-kube-api-access-wcdzv\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.928916 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:06 crc kubenswrapper[4960]: I1002 07:36:06.962478 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.030542 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.030622 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.030647 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a94097fa-b407-419c-9fad-5078a1f8967f-logs\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.030740 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-config-data\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.030765 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcdzv\" (UniqueName: \"kubernetes.io/projected/a94097fa-b407-419c-9fad-5078a1f8967f-kube-api-access-wcdzv\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.031602 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a94097fa-b407-419c-9fad-5078a1f8967f-logs\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.035415 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.035952 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.036686 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a94097fa-b407-419c-9fad-5078a1f8967f-config-data\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.052673 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcdzv\" (UniqueName: \"kubernetes.io/projected/a94097fa-b407-419c-9fad-5078a1f8967f-kube-api-access-wcdzv\") pod \"nova-metadata-0\" (UID: \"a94097fa-b407-419c-9fad-5078a1f8967f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.209431 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.437566 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.543899 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"189dd145-b9b7-4dc5-9fae-a44b359f8748","Type":"ContainerStarted","Data":"8e312b97152403fc023b9796742fee9f5385774c260cdd2188b4e04d9d1ca854"} Oct 02 07:36:07 crc kubenswrapper[4960]: I1002 07:36:07.678124 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.208669 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.344003 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d29d929-5ba0-44eb-9604-ef811016b504" path="/var/lib/kubelet/pods/0d29d929-5ba0-44eb-9604-ef811016b504/volumes" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.344605 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20" path="/var/lib/kubelet/pods/f173520c-e49b-4ea7-b9c5-1b5fa7ca9a20/volumes" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.355913 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-internal-tls-certs\") pod \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.356018 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-public-tls-certs\") pod \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.356126 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-logs\") pod \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.356213 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-combined-ca-bundle\") pod \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.356307 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx8gx\" (UniqueName: \"kubernetes.io/projected/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-kube-api-access-wx8gx\") pod \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.356382 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-config-data\") pod \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\" (UID: \"a968da5e-fbe3-4c48-a71f-c1b17d8731d8\") " Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.357998 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-logs" (OuterVolumeSpecName: "logs") pod "a968da5e-fbe3-4c48-a71f-c1b17d8731d8" (UID: "a968da5e-fbe3-4c48-a71f-c1b17d8731d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.363868 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-kube-api-access-wx8gx" (OuterVolumeSpecName: "kube-api-access-wx8gx") pod "a968da5e-fbe3-4c48-a71f-c1b17d8731d8" (UID: "a968da5e-fbe3-4c48-a71f-c1b17d8731d8"). InnerVolumeSpecName "kube-api-access-wx8gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.390882 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a968da5e-fbe3-4c48-a71f-c1b17d8731d8" (UID: "a968da5e-fbe3-4c48-a71f-c1b17d8731d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.400419 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-config-data" (OuterVolumeSpecName: "config-data") pod "a968da5e-fbe3-4c48-a71f-c1b17d8731d8" (UID: "a968da5e-fbe3-4c48-a71f-c1b17d8731d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.420504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a968da5e-fbe3-4c48-a71f-c1b17d8731d8" (UID: "a968da5e-fbe3-4c48-a71f-c1b17d8731d8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.438906 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a968da5e-fbe3-4c48-a71f-c1b17d8731d8" (UID: "a968da5e-fbe3-4c48-a71f-c1b17d8731d8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.459490 4960 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.459531 4960 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.459561 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.459573 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.459583 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx8gx\" (UniqueName: \"kubernetes.io/projected/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-kube-api-access-wx8gx\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.459595 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a968da5e-fbe3-4c48-a71f-c1b17d8731d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.559803 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a94097fa-b407-419c-9fad-5078a1f8967f","Type":"ContainerStarted","Data":"929a76f0c739ea98925809e74cb9998d4e2af55ee535c07f2faeed4417c50927"} Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.560512 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a94097fa-b407-419c-9fad-5078a1f8967f","Type":"ContainerStarted","Data":"1ce6c641b3277537e9d6ba138d63b16c2234346f9d42616fde8f7ea1f8ca706b"} Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.560559 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a94097fa-b407-419c-9fad-5078a1f8967f","Type":"ContainerStarted","Data":"4759ca569cfef80d0119fed7d172cc68ddb432f0e867d2bd8d3588f8b1cb01ce"} Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.566474 4960 generic.go:334] "Generic (PLEG): container finished" podID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerID="f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3" exitCode=0 Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.566549 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a968da5e-fbe3-4c48-a71f-c1b17d8731d8","Type":"ContainerDied","Data":"f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3"} Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.566582 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a968da5e-fbe3-4c48-a71f-c1b17d8731d8","Type":"ContainerDied","Data":"af1acfe163a05d6032619e78d8a3c3f009f0f596d7f34a882cd6ea2e569a62c4"} Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.566602 4960 scope.go:117] "RemoveContainer" containerID="f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.566742 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.581834 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"189dd145-b9b7-4dc5-9fae-a44b359f8748","Type":"ContainerStarted","Data":"2dbe91cc2dd37516bf43a8277d6c3222984e33a10088ff881556ddae4f47bc4c"} Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.596644 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.596615118 podStartE2EDuration="2.596615118s" podCreationTimestamp="2025-10-02 07:36:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:08.590958769 +0000 UTC m=+1189.622905076" watchObservedRunningTime="2025-10-02 07:36:08.596615118 +0000 UTC m=+1189.628561415" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.618391 4960 scope.go:117] "RemoveContainer" containerID="c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.625551 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.625523286 podStartE2EDuration="2.625523286s" podCreationTimestamp="2025-10-02 07:36:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:08.615125299 +0000 UTC m=+1189.647071586" watchObservedRunningTime="2025-10-02 07:36:08.625523286 +0000 UTC m=+1189.657469573" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.714639 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.723357 4960 scope.go:117] "RemoveContainer" containerID="f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3" Oct 02 07:36:08 crc kubenswrapper[4960]: E1002 07:36:08.730076 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3\": container with ID starting with f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3 not found: ID does not exist" containerID="f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.730322 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3"} err="failed to get container status \"f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3\": rpc error: code = NotFound desc = could not find container \"f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3\": container with ID starting with f81fb2c1c065a72674587812bf8d0b4faed1b199dbccac9ef640392437d600b3 not found: ID does not exist" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.730361 4960 scope.go:117] "RemoveContainer" containerID="c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913" Oct 02 07:36:08 crc kubenswrapper[4960]: E1002 07:36:08.733027 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913\": container with ID starting with c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913 not found: ID does not exist" containerID="c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.733062 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913"} err="failed to get container status \"c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913\": rpc error: code = NotFound desc = could not find container \"c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913\": container with ID starting with c8a98cdf2ad5ade8a1a14e28d9d885ed424184dda51e7f5adcefa239b47d9913 not found: ID does not exist" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.742078 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.752636 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:08 crc kubenswrapper[4960]: E1002 07:36:08.753242 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-log" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.753266 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-log" Oct 02 07:36:08 crc kubenswrapper[4960]: E1002 07:36:08.753317 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-api" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.753324 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-api" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.753511 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-api" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.753545 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" containerName="nova-api-log" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.754755 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.758242 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.758651 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.759436 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.762654 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.900648 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l48gl\" (UniqueName: \"kubernetes.io/projected/d889074b-a08e-49cf-8272-2b51fd6a9286-kube-api-access-l48gl\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.900751 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d889074b-a08e-49cf-8272-2b51fd6a9286-logs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.900830 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.900954 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-config-data\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.901053 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:08 crc kubenswrapper[4960]: I1002 07:36:08.901183 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-public-tls-certs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.003081 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-config-data\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.004356 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.004432 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-public-tls-certs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.004502 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l48gl\" (UniqueName: \"kubernetes.io/projected/d889074b-a08e-49cf-8272-2b51fd6a9286-kube-api-access-l48gl\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.004542 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d889074b-a08e-49cf-8272-2b51fd6a9286-logs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.004581 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.006094 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d889074b-a08e-49cf-8272-2b51fd6a9286-logs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.010366 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-public-tls-certs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.011073 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.012652 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.023338 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d889074b-a08e-49cf-8272-2b51fd6a9286-config-data\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.023665 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l48gl\" (UniqueName: \"kubernetes.io/projected/d889074b-a08e-49cf-8272-2b51fd6a9286-kube-api-access-l48gl\") pod \"nova-api-0\" (UID: \"d889074b-a08e-49cf-8272-2b51fd6a9286\") " pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.082315 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.573668 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:09 crc kubenswrapper[4960]: I1002 07:36:09.614779 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d889074b-a08e-49cf-8272-2b51fd6a9286","Type":"ContainerStarted","Data":"c677d49f09d6b9b4f164f7520467fc32693d28a05b0b902a39948551c074e7e8"} Oct 02 07:36:10 crc kubenswrapper[4960]: I1002 07:36:10.351323 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a968da5e-fbe3-4c48-a71f-c1b17d8731d8" path="/var/lib/kubelet/pods/a968da5e-fbe3-4c48-a71f-c1b17d8731d8/volumes" Oct 02 07:36:10 crc kubenswrapper[4960]: I1002 07:36:10.628850 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d889074b-a08e-49cf-8272-2b51fd6a9286","Type":"ContainerStarted","Data":"0f52c117fe920df1b91b7ce412d956e0410f145c53e08a00b3267bfa689c8adc"} Oct 02 07:36:10 crc kubenswrapper[4960]: I1002 07:36:10.628926 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d889074b-a08e-49cf-8272-2b51fd6a9286","Type":"ContainerStarted","Data":"7b905635d4efd8d6d228764f04c68bfa8ffd3db65070a1a4b3b33d8831f1ef73"} Oct 02 07:36:10 crc kubenswrapper[4960]: I1002 07:36:10.656398 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.656373288 podStartE2EDuration="2.656373288s" podCreationTimestamp="2025-10-02 07:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:10.650868972 +0000 UTC m=+1191.682815269" watchObservedRunningTime="2025-10-02 07:36:10.656373288 +0000 UTC m=+1191.688319595" Oct 02 07:36:11 crc kubenswrapper[4960]: I1002 07:36:11.963201 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 07:36:12 crc kubenswrapper[4960]: I1002 07:36:12.209778 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:36:12 crc kubenswrapper[4960]: I1002 07:36:12.209885 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:36:16 crc kubenswrapper[4960]: I1002 07:36:16.963593 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 07:36:16 crc kubenswrapper[4960]: I1002 07:36:16.994095 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 07:36:17 crc kubenswrapper[4960]: I1002 07:36:17.209956 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:36:17 crc kubenswrapper[4960]: I1002 07:36:17.210671 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:36:17 crc kubenswrapper[4960]: I1002 07:36:17.757641 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 07:36:18 crc kubenswrapper[4960]: I1002 07:36:18.219340 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a94097fa-b407-419c-9fad-5078a1f8967f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:18 crc kubenswrapper[4960]: I1002 07:36:18.226348 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a94097fa-b407-419c-9fad-5078a1f8967f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:19 crc kubenswrapper[4960]: I1002 07:36:19.083377 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:36:19 crc kubenswrapper[4960]: I1002 07:36:19.083450 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:36:20 crc kubenswrapper[4960]: I1002 07:36:20.101168 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d889074b-a08e-49cf-8272-2b51fd6a9286" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:20 crc kubenswrapper[4960]: I1002 07:36:20.101199 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d889074b-a08e-49cf-8272-2b51fd6a9286" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:21 crc kubenswrapper[4960]: I1002 07:36:21.001953 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 07:36:27 crc kubenswrapper[4960]: I1002 07:36:27.216034 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:36:27 crc kubenswrapper[4960]: I1002 07:36:27.221609 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:36:27 crc kubenswrapper[4960]: I1002 07:36:27.223654 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:36:27 crc kubenswrapper[4960]: I1002 07:36:27.866394 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:36:29 crc kubenswrapper[4960]: I1002 07:36:29.091289 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:36:29 crc kubenswrapper[4960]: I1002 07:36:29.092646 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:36:29 crc kubenswrapper[4960]: I1002 07:36:29.093109 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:36:29 crc kubenswrapper[4960]: I1002 07:36:29.093183 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:36:29 crc kubenswrapper[4960]: I1002 07:36:29.100750 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:36:29 crc kubenswrapper[4960]: I1002 07:36:29.102286 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:36:37 crc kubenswrapper[4960]: I1002 07:36:37.126244 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:36:38 crc kubenswrapper[4960]: I1002 07:36:38.793044 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:36:42 crc kubenswrapper[4960]: I1002 07:36:42.729052 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" containerName="rabbitmq" containerID="cri-o://af99fe471b99be3768d7cf9a79134ba1b0f4bcfc6f1b4ed446a107e137e07773" gracePeriod=604795 Oct 02 07:36:43 crc kubenswrapper[4960]: I1002 07:36:43.846774 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerName="rabbitmq" containerID="cri-o://9bd900f71f7c49f9c12c414640951d786bf29f7c016013398508dac1dd36e0e2" gracePeriod=604795 Oct 02 07:36:46 crc kubenswrapper[4960]: I1002 07:36:46.206573 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 02 07:36:46 crc kubenswrapper[4960]: I1002 07:36:46.538271 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.090460 4960 generic.go:334] "Generic (PLEG): container finished" podID="36c68533-435a-4bb7-b338-4cd484b155f6" containerID="af99fe471b99be3768d7cf9a79134ba1b0f4bcfc6f1b4ed446a107e137e07773" exitCode=0 Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.090563 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"36c68533-435a-4bb7-b338-4cd484b155f6","Type":"ContainerDied","Data":"af99fe471b99be3768d7cf9a79134ba1b0f4bcfc6f1b4ed446a107e137e07773"} Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.375397 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438065 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36c68533-435a-4bb7-b338-4cd484b155f6-pod-info\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438132 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc5zm\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-kube-api-access-xc5zm\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438189 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-plugins\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438290 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438414 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-config-data\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438527 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-erlang-cookie\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438579 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36c68533-435a-4bb7-b338-4cd484b155f6-erlang-cookie-secret\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438607 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-plugins-conf\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438632 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-tls\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438703 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-server-conf\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.438746 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-confd\") pod \"36c68533-435a-4bb7-b338-4cd484b155f6\" (UID: \"36c68533-435a-4bb7-b338-4cd484b155f6\") " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.441401 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.444806 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.446927 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.456070 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c68533-435a-4bb7-b338-4cd484b155f6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.456225 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.456306 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.456398 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/36c68533-435a-4bb7-b338-4cd484b155f6-pod-info" (OuterVolumeSpecName: "pod-info") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.462106 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-kube-api-access-xc5zm" (OuterVolumeSpecName: "kube-api-access-xc5zm") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "kube-api-access-xc5zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.488653 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-config-data" (OuterVolumeSpecName: "config-data") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.535037 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-server-conf" (OuterVolumeSpecName: "server-conf") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541079 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541111 4960 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36c68533-435a-4bb7-b338-4cd484b155f6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541122 4960 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541132 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541141 4960 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541149 4960 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36c68533-435a-4bb7-b338-4cd484b155f6-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541159 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc5zm\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-kube-api-access-xc5zm\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541168 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541199 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.541210 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36c68533-435a-4bb7-b338-4cd484b155f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.572962 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.624383 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "36c68533-435a-4bb7-b338-4cd484b155f6" (UID: "36c68533-435a-4bb7-b338-4cd484b155f6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.643183 4960 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4960]: I1002 07:36:49.643245 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36c68533-435a-4bb7-b338-4cd484b155f6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.109158 4960 generic.go:334] "Generic (PLEG): container finished" podID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerID="9bd900f71f7c49f9c12c414640951d786bf29f7c016013398508dac1dd36e0e2" exitCode=0 Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.109266 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8c7a2f9-57bf-4520-b078-ce75a2e051ad","Type":"ContainerDied","Data":"9bd900f71f7c49f9c12c414640951d786bf29f7c016013398508dac1dd36e0e2"} Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.112494 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"36c68533-435a-4bb7-b338-4cd484b155f6","Type":"ContainerDied","Data":"a1fffe3454e9faced5855e44fbec91b55d8e3a9481801baa93f7bac98a847031"} Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.112533 4960 scope.go:117] "RemoveContainer" containerID="af99fe471b99be3768d7cf9a79134ba1b0f4bcfc6f1b4ed446a107e137e07773" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.112621 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.148309 4960 scope.go:117] "RemoveContainer" containerID="9bd820698bae00820e1521389d7e85c9f91579903f903da17b0bbb829e24c442" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.158245 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.166810 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.193137 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:36:50 crc kubenswrapper[4960]: E1002 07:36:50.193749 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" containerName="setup-container" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.193776 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" containerName="setup-container" Oct 02 07:36:50 crc kubenswrapper[4960]: E1002 07:36:50.193795 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" containerName="rabbitmq" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.193804 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" containerName="rabbitmq" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.199196 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" containerName="rabbitmq" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.200602 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.208905 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.213435 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7wvn9" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.214245 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.220280 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.220562 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.220911 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.226472 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.227281 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259478 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259626 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259668 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259696 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259749 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/917895a0-f990-4182-bc76-14bfddefa2a4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259810 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rssvk\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-kube-api-access-rssvk\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259856 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259884 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/917895a0-f990-4182-bc76-14bfddefa2a4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259917 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-config-data\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.259943 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.358935 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36c68533-435a-4bb7-b338-4cd484b155f6" path="/var/lib/kubelet/pods/36c68533-435a-4bb7-b338-4cd484b155f6/volumes" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361077 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-config-data\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361130 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361155 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361233 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361271 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361290 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361323 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/917895a0-f990-4182-bc76-14bfddefa2a4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361376 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361394 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rssvk\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-kube-api-access-rssvk\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361413 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.361435 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/917895a0-f990-4182-bc76-14bfddefa2a4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.366898 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-config-data\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.367539 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.368459 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/917895a0-f990-4182-bc76-14bfddefa2a4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.369055 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.369347 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.370502 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.378092 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.387750 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/917895a0-f990-4182-bc76-14bfddefa2a4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.388706 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/917895a0-f990-4182-bc76-14bfddefa2a4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.389278 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.434664 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rssvk\" (UniqueName: \"kubernetes.io/projected/917895a0-f990-4182-bc76-14bfddefa2a4-kube-api-access-rssvk\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.461793 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"917895a0-f990-4182-bc76-14bfddefa2a4\") " pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.552080 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.616472 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.677274 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-tls\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.677320 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-confd\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.683703 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.778824 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-erlang-cookie-secret\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.778922 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-plugins\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779062 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-plugins-conf\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779099 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-erlang-cookie\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779137 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-config-data\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779178 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779206 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlzrr\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-kube-api-access-zlzrr\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779269 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-pod-info\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779327 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-server-conf\") pod \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\" (UID: \"f8c7a2f9-57bf-4520-b078-ce75a2e051ad\") " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779374 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779830 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.779844 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.780060 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.781553 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.783208 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.784036 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.789436 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-kube-api-access-zlzrr" (OuterVolumeSpecName: "kube-api-access-zlzrr") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "kube-api-access-zlzrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.789652 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-pod-info" (OuterVolumeSpecName: "pod-info") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.808151 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-config-data" (OuterVolumeSpecName: "config-data") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.821277 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.829170 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-server-conf" (OuterVolumeSpecName: "server-conf") pod "f8c7a2f9-57bf-4520-b078-ce75a2e051ad" (UID: "f8c7a2f9-57bf-4520-b078-ce75a2e051ad"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.881891 4960 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.881942 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.881951 4960 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.881965 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.881992 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.882031 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.882042 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlzrr\" (UniqueName: \"kubernetes.io/projected/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-kube-api-access-zlzrr\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.882051 4960 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.882060 4960 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8c7a2f9-57bf-4520-b078-ce75a2e051ad-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.901706 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 02 07:36:50 crc kubenswrapper[4960]: I1002 07:36:50.984777 4960 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.123830 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.131629 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8c7a2f9-57bf-4520-b078-ce75a2e051ad","Type":"ContainerDied","Data":"979d1f7ec5b1271b4a68630ac106b23942e7e072c3ce1ed0ffc50da9990dab23"} Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.131701 4960 scope.go:117] "RemoveContainer" containerID="9bd900f71f7c49f9c12c414640951d786bf29f7c016013398508dac1dd36e0e2" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.131871 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.243778 4960 scope.go:117] "RemoveContainer" containerID="08856e4cd8439c757573e71be33fb3eca2f53d6f1f0a9f9e755f234506844f0f" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.275327 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.284009 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.307936 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:36:51 crc kubenswrapper[4960]: E1002 07:36:51.308599 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerName="setup-container" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.308630 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerName="setup-container" Oct 02 07:36:51 crc kubenswrapper[4960]: E1002 07:36:51.308644 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerName="rabbitmq" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.308652 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerName="rabbitmq" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.308889 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" containerName="rabbitmq" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.312817 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.319484 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.319995 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.320589 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.320647 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.321138 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.321143 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.321357 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-s6vnf" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.330140 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.511683 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.512148 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.512288 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.512508 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbp7l\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-kube-api-access-tbp7l\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.512720 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.512890 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.513104 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.513266 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.513425 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43e28122-07de-46d8-b2bd-918405699fd7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.513589 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.513816 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43e28122-07de-46d8-b2bd-918405699fd7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.616364 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.616867 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.617055 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.617166 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.617288 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43e28122-07de-46d8-b2bd-918405699fd7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.617409 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.617616 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43e28122-07de-46d8-b2bd-918405699fd7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.617744 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.618061 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.618537 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.618664 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.618704 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.618744 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbp7l\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-kube-api-access-tbp7l\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.618761 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.618767 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.619069 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.623944 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43e28122-07de-46d8-b2bd-918405699fd7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.629089 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.629623 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43e28122-07de-46d8-b2bd-918405699fd7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.630337 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.631141 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43e28122-07de-46d8-b2bd-918405699fd7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.643395 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbp7l\" (UniqueName: \"kubernetes.io/projected/43e28122-07de-46d8-b2bd-918405699fd7-kube-api-access-tbp7l\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.660626 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43e28122-07de-46d8-b2bd-918405699fd7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:51 crc kubenswrapper[4960]: I1002 07:36:51.669335 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.153627 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"917895a0-f990-4182-bc76-14bfddefa2a4","Type":"ContainerStarted","Data":"c86c18db8852448a9925f79752974cca13d93b60795133a7ad77fc94c510de1b"} Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.190526 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.342550 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c7a2f9-57bf-4520-b078-ce75a2e051ad" path="/var/lib/kubelet/pods/f8c7a2f9-57bf-4520-b078-ce75a2e051ad/volumes" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.614652 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567b675d5f-zpp97"] Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.616917 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.619488 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.635304 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567b675d5f-zpp97"] Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.652800 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-nb\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.652868 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-config\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.652893 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-dns-svc\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.652949 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-sb\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.653013 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swfbz\" (UniqueName: \"kubernetes.io/projected/938c1c89-d930-4aac-ad02-ed489d74c0d6-kube-api-access-swfbz\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.653069 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-openstack-edpm-ipam\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.755849 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-sb\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.755970 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swfbz\" (UniqueName: \"kubernetes.io/projected/938c1c89-d930-4aac-ad02-ed489d74c0d6-kube-api-access-swfbz\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.756114 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-openstack-edpm-ipam\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.756911 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-sb\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.757412 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-openstack-edpm-ipam\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.757674 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-nb\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.758542 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-nb\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.759433 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-config\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.758650 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-config\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.759535 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-dns-svc\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.760352 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-dns-svc\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.779955 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swfbz\" (UniqueName: \"kubernetes.io/projected/938c1c89-d930-4aac-ad02-ed489d74c0d6-kube-api-access-swfbz\") pod \"dnsmasq-dns-567b675d5f-zpp97\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:52 crc kubenswrapper[4960]: I1002 07:36:52.964815 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:36:53 crc kubenswrapper[4960]: I1002 07:36:53.183852 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43e28122-07de-46d8-b2bd-918405699fd7","Type":"ContainerStarted","Data":"b3509143d75037c4ba96f57ee9922c6576f9a76847a9c1dc410633e17413fd25"} Oct 02 07:36:53 crc kubenswrapper[4960]: I1002 07:36:53.186484 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"917895a0-f990-4182-bc76-14bfddefa2a4","Type":"ContainerStarted","Data":"a81107499c88733816df16ff6f622207e5c931db683be3f2241dabf9b8f4ecf4"} Oct 02 07:36:53 crc kubenswrapper[4960]: I1002 07:36:53.454153 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567b675d5f-zpp97"] Oct 02 07:36:53 crc kubenswrapper[4960]: W1002 07:36:53.461921 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod938c1c89_d930_4aac_ad02_ed489d74c0d6.slice/crio-966554772e8ccc26a4c86416f315d4ad0ab4218da6d3aa371bf17fa52b6ddad0 WatchSource:0}: Error finding container 966554772e8ccc26a4c86416f315d4ad0ab4218da6d3aa371bf17fa52b6ddad0: Status 404 returned error can't find the container with id 966554772e8ccc26a4c86416f315d4ad0ab4218da6d3aa371bf17fa52b6ddad0 Oct 02 07:36:54 crc kubenswrapper[4960]: I1002 07:36:54.201944 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43e28122-07de-46d8-b2bd-918405699fd7","Type":"ContainerStarted","Data":"3f10f46df6780d6c9378232c0b0d0710ac412aafd552c583c0073daa761a1642"} Oct 02 07:36:54 crc kubenswrapper[4960]: I1002 07:36:54.203575 4960 generic.go:334] "Generic (PLEG): container finished" podID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerID="27c87bc22754f723bc0541a62c52534cf4db41187ec59fb8103d587299e78109" exitCode=0 Oct 02 07:36:54 crc kubenswrapper[4960]: I1002 07:36:54.203639 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" event={"ID":"938c1c89-d930-4aac-ad02-ed489d74c0d6","Type":"ContainerDied","Data":"27c87bc22754f723bc0541a62c52534cf4db41187ec59fb8103d587299e78109"} Oct 02 07:36:54 crc kubenswrapper[4960]: I1002 07:36:54.203688 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" event={"ID":"938c1c89-d930-4aac-ad02-ed489d74c0d6","Type":"ContainerStarted","Data":"966554772e8ccc26a4c86416f315d4ad0ab4218da6d3aa371bf17fa52b6ddad0"} Oct 02 07:36:55 crc kubenswrapper[4960]: I1002 07:36:55.215988 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" event={"ID":"938c1c89-d930-4aac-ad02-ed489d74c0d6","Type":"ContainerStarted","Data":"74ed78603c3047b838829ff023dac98a3c34de83b768bc43a64bb945810152e0"} Oct 02 07:36:55 crc kubenswrapper[4960]: I1002 07:36:55.248764 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" podStartSLOduration=3.248727392 podStartE2EDuration="3.248727392s" podCreationTimestamp="2025-10-02 07:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:55.235906894 +0000 UTC m=+1236.267853191" watchObservedRunningTime="2025-10-02 07:36:55.248727392 +0000 UTC m=+1236.280673699" Oct 02 07:36:56 crc kubenswrapper[4960]: I1002 07:36:56.228026 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:37:02 crc kubenswrapper[4960]: I1002 07:37:02.967498 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.045537 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54bd68d8df-bhlzv"] Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.045860 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" podUID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerName="dnsmasq-dns" containerID="cri-o://bed36969808958469052a09c93af5d4dc9fc02d7e427a6278141ed81476e2882" gracePeriod=10 Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.321699 4960 generic.go:334] "Generic (PLEG): container finished" podID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerID="bed36969808958469052a09c93af5d4dc9fc02d7e427a6278141ed81476e2882" exitCode=0 Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.322232 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" event={"ID":"aed2fbca-b0bd-4304-9401-4b4257d33c82","Type":"ContainerDied","Data":"bed36969808958469052a09c93af5d4dc9fc02d7e427a6278141ed81476e2882"} Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.325738 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b897d8569-xzwfc"] Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.328370 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.344149 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b897d8569-xzwfc"] Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.406181 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdpfs\" (UniqueName: \"kubernetes.io/projected/17b9b7cb-39cd-45fd-9653-fe88daede51a-kube-api-access-pdpfs\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.406560 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-config\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.406644 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-nb\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.406691 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-dns-svc\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.406738 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.406768 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-sb\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.509514 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.509572 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-sb\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.509685 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdpfs\" (UniqueName: \"kubernetes.io/projected/17b9b7cb-39cd-45fd-9653-fe88daede51a-kube-api-access-pdpfs\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.509928 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-config\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.510033 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-nb\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.510088 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-dns-svc\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.510521 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.510829 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-config\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.510953 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-dns-svc\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.511019 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-nb\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.512161 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-sb\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.535546 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdpfs\" (UniqueName: \"kubernetes.io/projected/17b9b7cb-39cd-45fd-9653-fe88daede51a-kube-api-access-pdpfs\") pod \"dnsmasq-dns-5b897d8569-xzwfc\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.635764 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.654108 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.715249 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-nb\") pod \"aed2fbca-b0bd-4304-9401-4b4257d33c82\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.717161 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-dns-svc\") pod \"aed2fbca-b0bd-4304-9401-4b4257d33c82\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.717643 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-sb\") pod \"aed2fbca-b0bd-4304-9401-4b4257d33c82\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.717758 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n847s\" (UniqueName: \"kubernetes.io/projected/aed2fbca-b0bd-4304-9401-4b4257d33c82-kube-api-access-n847s\") pod \"aed2fbca-b0bd-4304-9401-4b4257d33c82\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.718031 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-config\") pod \"aed2fbca-b0bd-4304-9401-4b4257d33c82\" (UID: \"aed2fbca-b0bd-4304-9401-4b4257d33c82\") " Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.740466 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed2fbca-b0bd-4304-9401-4b4257d33c82-kube-api-access-n847s" (OuterVolumeSpecName: "kube-api-access-n847s") pod "aed2fbca-b0bd-4304-9401-4b4257d33c82" (UID: "aed2fbca-b0bd-4304-9401-4b4257d33c82"). InnerVolumeSpecName "kube-api-access-n847s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.794560 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aed2fbca-b0bd-4304-9401-4b4257d33c82" (UID: "aed2fbca-b0bd-4304-9401-4b4257d33c82"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.797904 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aed2fbca-b0bd-4304-9401-4b4257d33c82" (UID: "aed2fbca-b0bd-4304-9401-4b4257d33c82"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.798300 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-config" (OuterVolumeSpecName: "config") pod "aed2fbca-b0bd-4304-9401-4b4257d33c82" (UID: "aed2fbca-b0bd-4304-9401-4b4257d33c82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.807597 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aed2fbca-b0bd-4304-9401-4b4257d33c82" (UID: "aed2fbca-b0bd-4304-9401-4b4257d33c82"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.823652 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.824218 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.824309 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.824368 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed2fbca-b0bd-4304-9401-4b4257d33c82-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:03 crc kubenswrapper[4960]: I1002 07:37:03.824437 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n847s\" (UniqueName: \"kubernetes.io/projected/aed2fbca-b0bd-4304-9401-4b4257d33c82-kube-api-access-n847s\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.204919 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b897d8569-xzwfc"] Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.408295 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" event={"ID":"17b9b7cb-39cd-45fd-9653-fe88daede51a","Type":"ContainerStarted","Data":"2bb83c30931c6c2260b1b99b1d53b6f5158b1b65725b1c94dbc1aab0fe8a11ac"} Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.408371 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6"] Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.400769 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" Oct 02 07:37:04 crc kubenswrapper[4960]: E1002 07:37:04.411143 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerName="init" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.411183 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerName="init" Oct 02 07:37:04 crc kubenswrapper[4960]: E1002 07:37:04.411262 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerName="dnsmasq-dns" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.411270 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerName="dnsmasq-dns" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.411708 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed2fbca-b0bd-4304-9401-4b4257d33c82" containerName="dnsmasq-dns" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.413682 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6"] Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.413840 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bd68d8df-bhlzv" event={"ID":"aed2fbca-b0bd-4304-9401-4b4257d33c82","Type":"ContainerDied","Data":"0e35a0aeccb19160db3b6dc6cc8a314605ca2568ad96b30eedcbc20ad2c3924d"} Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.414079 4960 scope.go:117] "RemoveContainer" containerID="bed36969808958469052a09c93af5d4dc9fc02d7e427a6278141ed81476e2882" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.414079 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.420891 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.421428 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.421782 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.422455 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.456526 4960 scope.go:117] "RemoveContainer" containerID="0903ea7d79f98e51cafbc9b72c96bbe58ac4986759799d34ccf926f41dd93261" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.482800 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54bd68d8df-bhlzv"] Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.490715 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54bd68d8df-bhlzv"] Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.571319 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.571787 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.571955 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c96gr\" (UniqueName: \"kubernetes.io/projected/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-kube-api-access-c96gr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.572380 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.675578 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.675646 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c96gr\" (UniqueName: \"kubernetes.io/projected/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-kube-api-access-c96gr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.675716 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.675868 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.681161 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.681212 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.681523 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.695548 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c96gr\" (UniqueName: \"kubernetes.io/projected/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-kube-api-access-c96gr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:04 crc kubenswrapper[4960]: I1002 07:37:04.752012 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:05 crc kubenswrapper[4960]: I1002 07:37:05.306061 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6"] Oct 02 07:37:05 crc kubenswrapper[4960]: I1002 07:37:05.314191 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:37:05 crc kubenswrapper[4960]: I1002 07:37:05.411959 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" event={"ID":"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05","Type":"ContainerStarted","Data":"ab146d02e9ebbd99abf342fe70621b15674a3aeef93c66cb1e0cc61548ef5a32"} Oct 02 07:37:05 crc kubenswrapper[4960]: I1002 07:37:05.413939 4960 generic.go:334] "Generic (PLEG): container finished" podID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerID="2606b40635861e61bd9ea79b4756136ec706e29498af737a07d7d3d090bfc0ed" exitCode=0 Oct 02 07:37:05 crc kubenswrapper[4960]: I1002 07:37:05.414080 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" event={"ID":"17b9b7cb-39cd-45fd-9653-fe88daede51a","Type":"ContainerDied","Data":"2606b40635861e61bd9ea79b4756136ec706e29498af737a07d7d3d090bfc0ed"} Oct 02 07:37:06 crc kubenswrapper[4960]: I1002 07:37:06.347154 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed2fbca-b0bd-4304-9401-4b4257d33c82" path="/var/lib/kubelet/pods/aed2fbca-b0bd-4304-9401-4b4257d33c82/volumes" Oct 02 07:37:06 crc kubenswrapper[4960]: I1002 07:37:06.435254 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" event={"ID":"17b9b7cb-39cd-45fd-9653-fe88daede51a","Type":"ContainerStarted","Data":"94627eaaf9bb69e5565cec1e0d2fdf8a94a0592615e83f84adb2c84da8f1a2cf"} Oct 02 07:37:06 crc kubenswrapper[4960]: I1002 07:37:06.435618 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:06 crc kubenswrapper[4960]: I1002 07:37:06.468642 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" podStartSLOduration=3.468601032 podStartE2EDuration="3.468601032s" podCreationTimestamp="2025-10-02 07:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:06.453781041 +0000 UTC m=+1247.485727328" watchObservedRunningTime="2025-10-02 07:37:06.468601032 +0000 UTC m=+1247.500547349" Oct 02 07:37:13 crc kubenswrapper[4960]: I1002 07:37:13.655856 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 07:37:13 crc kubenswrapper[4960]: I1002 07:37:13.760221 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567b675d5f-zpp97"] Oct 02 07:37:13 crc kubenswrapper[4960]: I1002 07:37:13.761069 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" podUID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerName="dnsmasq-dns" containerID="cri-o://74ed78603c3047b838829ff023dac98a3c34de83b768bc43a64bb945810152e0" gracePeriod=10 Oct 02 07:37:14 crc kubenswrapper[4960]: I1002 07:37:14.551708 4960 generic.go:334] "Generic (PLEG): container finished" podID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerID="74ed78603c3047b838829ff023dac98a3c34de83b768bc43a64bb945810152e0" exitCode=0 Oct 02 07:37:14 crc kubenswrapper[4960]: I1002 07:37:14.551747 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" event={"ID":"938c1c89-d930-4aac-ad02-ed489d74c0d6","Type":"ContainerDied","Data":"74ed78603c3047b838829ff023dac98a3c34de83b768bc43a64bb945810152e0"} Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.146891 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.273879 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-sb\") pod \"938c1c89-d930-4aac-ad02-ed489d74c0d6\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.274016 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-openstack-edpm-ipam\") pod \"938c1c89-d930-4aac-ad02-ed489d74c0d6\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.274067 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-nb\") pod \"938c1c89-d930-4aac-ad02-ed489d74c0d6\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.274158 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swfbz\" (UniqueName: \"kubernetes.io/projected/938c1c89-d930-4aac-ad02-ed489d74c0d6-kube-api-access-swfbz\") pod \"938c1c89-d930-4aac-ad02-ed489d74c0d6\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.274205 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-dns-svc\") pod \"938c1c89-d930-4aac-ad02-ed489d74c0d6\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.274272 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-config\") pod \"938c1c89-d930-4aac-ad02-ed489d74c0d6\" (UID: \"938c1c89-d930-4aac-ad02-ed489d74c0d6\") " Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.279271 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/938c1c89-d930-4aac-ad02-ed489d74c0d6-kube-api-access-swfbz" (OuterVolumeSpecName: "kube-api-access-swfbz") pod "938c1c89-d930-4aac-ad02-ed489d74c0d6" (UID: "938c1c89-d930-4aac-ad02-ed489d74c0d6"). InnerVolumeSpecName "kube-api-access-swfbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.326285 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-config" (OuterVolumeSpecName: "config") pod "938c1c89-d930-4aac-ad02-ed489d74c0d6" (UID: "938c1c89-d930-4aac-ad02-ed489d74c0d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.326300 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "938c1c89-d930-4aac-ad02-ed489d74c0d6" (UID: "938c1c89-d930-4aac-ad02-ed489d74c0d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.328026 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "938c1c89-d930-4aac-ad02-ed489d74c0d6" (UID: "938c1c89-d930-4aac-ad02-ed489d74c0d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.330596 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "938c1c89-d930-4aac-ad02-ed489d74c0d6" (UID: "938c1c89-d930-4aac-ad02-ed489d74c0d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.334505 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "938c1c89-d930-4aac-ad02-ed489d74c0d6" (UID: "938c1c89-d930-4aac-ad02-ed489d74c0d6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.376569 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swfbz\" (UniqueName: \"kubernetes.io/projected/938c1c89-d930-4aac-ad02-ed489d74c0d6-kube-api-access-swfbz\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.376603 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.376615 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.376624 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.376635 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.376643 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/938c1c89-d930-4aac-ad02-ed489d74c0d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.565627 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" event={"ID":"938c1c89-d930-4aac-ad02-ed489d74c0d6","Type":"ContainerDied","Data":"966554772e8ccc26a4c86416f315d4ad0ab4218da6d3aa371bf17fa52b6ddad0"} Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.565750 4960 scope.go:117] "RemoveContainer" containerID="74ed78603c3047b838829ff023dac98a3c34de83b768bc43a64bb945810152e0" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.565649 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567b675d5f-zpp97" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.573907 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" event={"ID":"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05","Type":"ContainerStarted","Data":"9da040311f074cb1188e386be5749e2a92e4195f811a6efcac29e6f5ed23f9f9"} Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.598351 4960 scope.go:117] "RemoveContainer" containerID="27c87bc22754f723bc0541a62c52534cf4db41187ec59fb8103d587299e78109" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.610073 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" podStartSLOduration=2.088949222 podStartE2EDuration="11.610039379s" podCreationTimestamp="2025-10-02 07:37:04 +0000 UTC" firstStartedPulling="2025-10-02 07:37:05.313816347 +0000 UTC m=+1246.345762634" lastFinishedPulling="2025-10-02 07:37:14.834906504 +0000 UTC m=+1255.866852791" observedRunningTime="2025-10-02 07:37:15.596167944 +0000 UTC m=+1256.628114231" watchObservedRunningTime="2025-10-02 07:37:15.610039379 +0000 UTC m=+1256.641985706" Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.637367 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567b675d5f-zpp97"] Oct 02 07:37:15 crc kubenswrapper[4960]: I1002 07:37:15.649219 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567b675d5f-zpp97"] Oct 02 07:37:16 crc kubenswrapper[4960]: I1002 07:37:16.343582 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="938c1c89-d930-4aac-ad02-ed489d74c0d6" path="/var/lib/kubelet/pods/938c1c89-d930-4aac-ad02-ed489d74c0d6/volumes" Oct 02 07:37:26 crc kubenswrapper[4960]: I1002 07:37:26.698356 4960 generic.go:334] "Generic (PLEG): container finished" podID="917895a0-f990-4182-bc76-14bfddefa2a4" containerID="a81107499c88733816df16ff6f622207e5c931db683be3f2241dabf9b8f4ecf4" exitCode=0 Oct 02 07:37:26 crc kubenswrapper[4960]: I1002 07:37:26.698515 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"917895a0-f990-4182-bc76-14bfddefa2a4","Type":"ContainerDied","Data":"a81107499c88733816df16ff6f622207e5c931db683be3f2241dabf9b8f4ecf4"} Oct 02 07:37:26 crc kubenswrapper[4960]: I1002 07:37:26.703389 4960 generic.go:334] "Generic (PLEG): container finished" podID="7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" containerID="9da040311f074cb1188e386be5749e2a92e4195f811a6efcac29e6f5ed23f9f9" exitCode=0 Oct 02 07:37:26 crc kubenswrapper[4960]: I1002 07:37:26.703437 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" event={"ID":"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05","Type":"ContainerDied","Data":"9da040311f074cb1188e386be5749e2a92e4195f811a6efcac29e6f5ed23f9f9"} Oct 02 07:37:26 crc kubenswrapper[4960]: I1002 07:37:26.706474 4960 generic.go:334] "Generic (PLEG): container finished" podID="43e28122-07de-46d8-b2bd-918405699fd7" containerID="3f10f46df6780d6c9378232c0b0d0710ac412aafd552c583c0073daa761a1642" exitCode=0 Oct 02 07:37:26 crc kubenswrapper[4960]: I1002 07:37:26.706531 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43e28122-07de-46d8-b2bd-918405699fd7","Type":"ContainerDied","Data":"3f10f46df6780d6c9378232c0b0d0710ac412aafd552c583c0073daa761a1642"} Oct 02 07:37:27 crc kubenswrapper[4960]: I1002 07:37:27.717931 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43e28122-07de-46d8-b2bd-918405699fd7","Type":"ContainerStarted","Data":"4235470d659524b0ab8dd2714e9bb8e3f3e84fc45f3f0f4eee053b5d750e0405"} Oct 02 07:37:27 crc kubenswrapper[4960]: I1002 07:37:27.718789 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:37:27 crc kubenswrapper[4960]: I1002 07:37:27.722488 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"917895a0-f990-4182-bc76-14bfddefa2a4","Type":"ContainerStarted","Data":"21014e100850575319777bbbd6ce0a61a5c7fd641c9b8894af020c7d2c3d491a"} Oct 02 07:37:27 crc kubenswrapper[4960]: I1002 07:37:27.722881 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 07:37:27 crc kubenswrapper[4960]: I1002 07:37:27.756819 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.756797144 podStartE2EDuration="36.756797144s" podCreationTimestamp="2025-10-02 07:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:27.747861283 +0000 UTC m=+1268.779807570" watchObservedRunningTime="2025-10-02 07:37:27.756797144 +0000 UTC m=+1268.788743431" Oct 02 07:37:27 crc kubenswrapper[4960]: I1002 07:37:27.803536 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.803502296 podStartE2EDuration="37.803502296s" podCreationTimestamp="2025-10-02 07:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:27.787186615 +0000 UTC m=+1268.819132982" watchObservedRunningTime="2025-10-02 07:37:27.803502296 +0000 UTC m=+1268.835448623" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.241722 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.288504 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-ssh-key\") pod \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.288568 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c96gr\" (UniqueName: \"kubernetes.io/projected/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-kube-api-access-c96gr\") pod \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.288638 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-repo-setup-combined-ca-bundle\") pod \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.288734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-inventory\") pod \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\" (UID: \"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05\") " Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.298880 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-kube-api-access-c96gr" (OuterVolumeSpecName: "kube-api-access-c96gr") pod "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" (UID: "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05"). InnerVolumeSpecName "kube-api-access-c96gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.305301 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" (UID: "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.336144 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-inventory" (OuterVolumeSpecName: "inventory") pod "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" (UID: "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.347775 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" (UID: "7dac1c0f-0383-4a34-a3b9-10dbd82dbd05"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.391343 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.391376 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c96gr\" (UniqueName: \"kubernetes.io/projected/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-kube-api-access-c96gr\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.391390 4960 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.391405 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.735178 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" event={"ID":"7dac1c0f-0383-4a34-a3b9-10dbd82dbd05","Type":"ContainerDied","Data":"ab146d02e9ebbd99abf342fe70621b15674a3aeef93c66cb1e0cc61548ef5a32"} Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.735229 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab146d02e9ebbd99abf342fe70621b15674a3aeef93c66cb1e0cc61548ef5a32" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.735323 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.860566 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk"] Oct 02 07:37:28 crc kubenswrapper[4960]: E1002 07:37:28.861079 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.861103 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:37:28 crc kubenswrapper[4960]: E1002 07:37:28.861126 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerName="dnsmasq-dns" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.861136 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerName="dnsmasq-dns" Oct 02 07:37:28 crc kubenswrapper[4960]: E1002 07:37:28.861155 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerName="init" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.861164 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerName="init" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.861366 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="938c1c89-d930-4aac-ad02-ed489d74c0d6" containerName="dnsmasq-dns" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.861395 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.862169 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.865031 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.865432 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.865565 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.865728 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.880539 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk"] Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.902589 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.902698 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.902774 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:28 crc kubenswrapper[4960]: I1002 07:37:28.902819 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7b4n\" (UniqueName: \"kubernetes.io/projected/881f2dfd-9294-444d-9c05-18b72a9128a2-kube-api-access-r7b4n\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.015665 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.015793 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.015853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7b4n\" (UniqueName: \"kubernetes.io/projected/881f2dfd-9294-444d-9c05-18b72a9128a2-kube-api-access-r7b4n\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.016478 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.021424 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.022955 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.023136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.037197 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7b4n\" (UniqueName: \"kubernetes.io/projected/881f2dfd-9294-444d-9c05-18b72a9128a2-kube-api-access-r7b4n\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.180608 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:37:29 crc kubenswrapper[4960]: I1002 07:37:29.786434 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk"] Oct 02 07:37:29 crc kubenswrapper[4960]: W1002 07:37:29.797971 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod881f2dfd_9294_444d_9c05_18b72a9128a2.slice/crio-5ee1145348db2c67a794b6c7dfcad767baecf87c9ee3459c9ad802efed130667 WatchSource:0}: Error finding container 5ee1145348db2c67a794b6c7dfcad767baecf87c9ee3459c9ad802efed130667: Status 404 returned error can't find the container with id 5ee1145348db2c67a794b6c7dfcad767baecf87c9ee3459c9ad802efed130667 Oct 02 07:37:30 crc kubenswrapper[4960]: I1002 07:37:30.762624 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" event={"ID":"881f2dfd-9294-444d-9c05-18b72a9128a2","Type":"ContainerStarted","Data":"c0b040b5f28ce1989e16094bd189d8de6a543e5af0f71f799568e35acad5be5e"} Oct 02 07:37:30 crc kubenswrapper[4960]: I1002 07:37:30.763832 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" event={"ID":"881f2dfd-9294-444d-9c05-18b72a9128a2","Type":"ContainerStarted","Data":"5ee1145348db2c67a794b6c7dfcad767baecf87c9ee3459c9ad802efed130667"} Oct 02 07:37:30 crc kubenswrapper[4960]: I1002 07:37:30.797577 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" podStartSLOduration=2.245131103 podStartE2EDuration="2.79754305s" podCreationTimestamp="2025-10-02 07:37:28 +0000 UTC" firstStartedPulling="2025-10-02 07:37:29.804181448 +0000 UTC m=+1270.836127735" lastFinishedPulling="2025-10-02 07:37:30.356593395 +0000 UTC m=+1271.388539682" observedRunningTime="2025-10-02 07:37:30.782766601 +0000 UTC m=+1271.814712898" watchObservedRunningTime="2025-10-02 07:37:30.79754305 +0000 UTC m=+1271.829489367" Oct 02 07:37:40 crc kubenswrapper[4960]: I1002 07:37:40.556302 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 07:37:41 crc kubenswrapper[4960]: I1002 07:37:41.674238 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:37:59 crc kubenswrapper[4960]: I1002 07:37:59.150076 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:37:59 crc kubenswrapper[4960]: I1002 07:37:59.150871 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:38:29 crc kubenswrapper[4960]: I1002 07:38:29.149934 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:38:29 crc kubenswrapper[4960]: I1002 07:38:29.150852 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:38:41 crc kubenswrapper[4960]: I1002 07:38:41.447809 4960 scope.go:117] "RemoveContainer" containerID="628286f25086ad1a428009e59f9e7361937a92068d53660e27afe48bf4ad4032" Oct 02 07:38:41 crc kubenswrapper[4960]: I1002 07:38:41.488720 4960 scope.go:117] "RemoveContainer" containerID="b4f74894f81174a81e36d60ae244835da47f87bc33c704fe9b5a9a4d469e71c3" Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.150404 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.151536 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.151643 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.153449 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5bb746b73018c3c6e2b85b664f4b1896310fbe4f05d3269e3eee3a9625ffe9d"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.153564 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://f5bb746b73018c3c6e2b85b664f4b1896310fbe4f05d3269e3eee3a9625ffe9d" gracePeriod=600 Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.808876 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="f5bb746b73018c3c6e2b85b664f4b1896310fbe4f05d3269e3eee3a9625ffe9d" exitCode=0 Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.808964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"f5bb746b73018c3c6e2b85b664f4b1896310fbe4f05d3269e3eee3a9625ffe9d"} Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.809562 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678"} Oct 02 07:38:59 crc kubenswrapper[4960]: I1002 07:38:59.809614 4960 scope.go:117] "RemoveContainer" containerID="ae97637d73736b75a47ea930cb3cf4b53e5a377c7969711a5fc9d3103045505c" Oct 02 07:39:41 crc kubenswrapper[4960]: I1002 07:39:41.650008 4960 scope.go:117] "RemoveContainer" containerID="d6a87fa707ede52eb6ec04e603acdde85d44b5c90e49497c2f4d6c8c805db2de" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.521313 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vrch8"] Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.526295 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.534443 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vrch8"] Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.558626 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-catalog-content\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.558920 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-utilities\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.558998 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rd97\" (UniqueName: \"kubernetes.io/projected/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-kube-api-access-9rd97\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.660781 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-utilities\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.660841 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rd97\" (UniqueName: \"kubernetes.io/projected/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-kube-api-access-9rd97\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.660884 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-catalog-content\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.661621 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-utilities\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.661652 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-catalog-content\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.693143 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rd97\" (UniqueName: \"kubernetes.io/projected/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-kube-api-access-9rd97\") pod \"redhat-marketplace-vrch8\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:12 crc kubenswrapper[4960]: I1002 07:40:12.859384 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:13 crc kubenswrapper[4960]: I1002 07:40:13.356401 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vrch8"] Oct 02 07:40:13 crc kubenswrapper[4960]: W1002 07:40:13.363532 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb5e91eb_0f69_43fe_9b3d_9a75608c170a.slice/crio-af483e6fa44e13a4391f9b5f6a4ff14ddd2210360adf88a1862dbb7729aad18e WatchSource:0}: Error finding container af483e6fa44e13a4391f9b5f6a4ff14ddd2210360adf88a1862dbb7729aad18e: Status 404 returned error can't find the container with id af483e6fa44e13a4391f9b5f6a4ff14ddd2210360adf88a1862dbb7729aad18e Oct 02 07:40:13 crc kubenswrapper[4960]: I1002 07:40:13.753176 4960 generic.go:334] "Generic (PLEG): container finished" podID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerID="d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760" exitCode=0 Oct 02 07:40:13 crc kubenswrapper[4960]: I1002 07:40:13.753244 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vrch8" event={"ID":"cb5e91eb-0f69-43fe-9b3d-9a75608c170a","Type":"ContainerDied","Data":"d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760"} Oct 02 07:40:13 crc kubenswrapper[4960]: I1002 07:40:13.753287 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vrch8" event={"ID":"cb5e91eb-0f69-43fe-9b3d-9a75608c170a","Type":"ContainerStarted","Data":"af483e6fa44e13a4391f9b5f6a4ff14ddd2210360adf88a1862dbb7729aad18e"} Oct 02 07:40:14 crc kubenswrapper[4960]: I1002 07:40:14.770899 4960 generic.go:334] "Generic (PLEG): container finished" podID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerID="ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5" exitCode=0 Oct 02 07:40:14 crc kubenswrapper[4960]: I1002 07:40:14.770969 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vrch8" event={"ID":"cb5e91eb-0f69-43fe-9b3d-9a75608c170a","Type":"ContainerDied","Data":"ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5"} Oct 02 07:40:15 crc kubenswrapper[4960]: I1002 07:40:15.789348 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vrch8" event={"ID":"cb5e91eb-0f69-43fe-9b3d-9a75608c170a","Type":"ContainerStarted","Data":"cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883"} Oct 02 07:40:15 crc kubenswrapper[4960]: I1002 07:40:15.821821 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vrch8" podStartSLOduration=2.389255644 podStartE2EDuration="3.821797213s" podCreationTimestamp="2025-10-02 07:40:12 +0000 UTC" firstStartedPulling="2025-10-02 07:40:13.755663589 +0000 UTC m=+1434.787609876" lastFinishedPulling="2025-10-02 07:40:15.188205118 +0000 UTC m=+1436.220151445" observedRunningTime="2025-10-02 07:40:15.816966077 +0000 UTC m=+1436.848912404" watchObservedRunningTime="2025-10-02 07:40:15.821797213 +0000 UTC m=+1436.853743510" Oct 02 07:40:22 crc kubenswrapper[4960]: I1002 07:40:22.859813 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:22 crc kubenswrapper[4960]: I1002 07:40:22.862162 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:22 crc kubenswrapper[4960]: I1002 07:40:22.927577 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:22 crc kubenswrapper[4960]: I1002 07:40:22.996311 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:23 crc kubenswrapper[4960]: I1002 07:40:23.175866 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vrch8"] Oct 02 07:40:24 crc kubenswrapper[4960]: I1002 07:40:24.905676 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vrch8" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="registry-server" containerID="cri-o://cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883" gracePeriod=2 Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.405570 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.571796 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-utilities\") pod \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.571965 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rd97\" (UniqueName: \"kubernetes.io/projected/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-kube-api-access-9rd97\") pod \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.572110 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-catalog-content\") pod \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\" (UID: \"cb5e91eb-0f69-43fe-9b3d-9a75608c170a\") " Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.573691 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-utilities" (OuterVolumeSpecName: "utilities") pod "cb5e91eb-0f69-43fe-9b3d-9a75608c170a" (UID: "cb5e91eb-0f69-43fe-9b3d-9a75608c170a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.585873 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-kube-api-access-9rd97" (OuterVolumeSpecName: "kube-api-access-9rd97") pod "cb5e91eb-0f69-43fe-9b3d-9a75608c170a" (UID: "cb5e91eb-0f69-43fe-9b3d-9a75608c170a"). InnerVolumeSpecName "kube-api-access-9rd97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.586814 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb5e91eb-0f69-43fe-9b3d-9a75608c170a" (UID: "cb5e91eb-0f69-43fe-9b3d-9a75608c170a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.675783 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.675820 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rd97\" (UniqueName: \"kubernetes.io/projected/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-kube-api-access-9rd97\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.675836 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5e91eb-0f69-43fe-9b3d-9a75608c170a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.921668 4960 generic.go:334] "Generic (PLEG): container finished" podID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerID="cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883" exitCode=0 Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.921730 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vrch8" event={"ID":"cb5e91eb-0f69-43fe-9b3d-9a75608c170a","Type":"ContainerDied","Data":"cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883"} Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.921771 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vrch8" event={"ID":"cb5e91eb-0f69-43fe-9b3d-9a75608c170a","Type":"ContainerDied","Data":"af483e6fa44e13a4391f9b5f6a4ff14ddd2210360adf88a1862dbb7729aad18e"} Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.921793 4960 scope.go:117] "RemoveContainer" containerID="cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.921808 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vrch8" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.956611 4960 scope.go:117] "RemoveContainer" containerID="ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5" Oct 02 07:40:25 crc kubenswrapper[4960]: I1002 07:40:25.998344 4960 scope.go:117] "RemoveContainer" containerID="d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760" Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.012850 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vrch8"] Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.022476 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vrch8"] Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.060428 4960 scope.go:117] "RemoveContainer" containerID="cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883" Oct 02 07:40:26 crc kubenswrapper[4960]: E1002 07:40:26.062314 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883\": container with ID starting with cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883 not found: ID does not exist" containerID="cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883" Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.062381 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883"} err="failed to get container status \"cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883\": rpc error: code = NotFound desc = could not find container \"cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883\": container with ID starting with cb869e86c824ecd8fe7292e7bed8232599251b69f1a79b3687e6fa569f4de883 not found: ID does not exist" Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.062420 4960 scope.go:117] "RemoveContainer" containerID="ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5" Oct 02 07:40:26 crc kubenswrapper[4960]: E1002 07:40:26.063111 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5\": container with ID starting with ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5 not found: ID does not exist" containerID="ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5" Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.063195 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5"} err="failed to get container status \"ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5\": rpc error: code = NotFound desc = could not find container \"ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5\": container with ID starting with ff23d86c02c8e76259c4df2e32cab282a196a5d227da8b688309d870678f3fc5 not found: ID does not exist" Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.063241 4960 scope.go:117] "RemoveContainer" containerID="d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760" Oct 02 07:40:26 crc kubenswrapper[4960]: E1002 07:40:26.063862 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760\": container with ID starting with d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760 not found: ID does not exist" containerID="d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760" Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.063905 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760"} err="failed to get container status \"d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760\": rpc error: code = NotFound desc = could not find container \"d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760\": container with ID starting with d2ad7d630827ac1f2b59265515dac51df56426bf2a84a1bf1f9ac77705c90760 not found: ID does not exist" Oct 02 07:40:26 crc kubenswrapper[4960]: I1002 07:40:26.351485 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" path="/var/lib/kubelet/pods/cb5e91eb-0f69-43fe-9b3d-9a75608c170a/volumes" Oct 02 07:40:37 crc kubenswrapper[4960]: I1002 07:40:37.080058 4960 generic.go:334] "Generic (PLEG): container finished" podID="881f2dfd-9294-444d-9c05-18b72a9128a2" containerID="c0b040b5f28ce1989e16094bd189d8de6a543e5af0f71f799568e35acad5be5e" exitCode=0 Oct 02 07:40:37 crc kubenswrapper[4960]: I1002 07:40:37.080388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" event={"ID":"881f2dfd-9294-444d-9c05-18b72a9128a2","Type":"ContainerDied","Data":"c0b040b5f28ce1989e16094bd189d8de6a543e5af0f71f799568e35acad5be5e"} Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.582233 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.706019 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-inventory\") pod \"881f2dfd-9294-444d-9c05-18b72a9128a2\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.706604 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-ssh-key\") pod \"881f2dfd-9294-444d-9c05-18b72a9128a2\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.706662 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-bootstrap-combined-ca-bundle\") pod \"881f2dfd-9294-444d-9c05-18b72a9128a2\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.706710 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7b4n\" (UniqueName: \"kubernetes.io/projected/881f2dfd-9294-444d-9c05-18b72a9128a2-kube-api-access-r7b4n\") pod \"881f2dfd-9294-444d-9c05-18b72a9128a2\" (UID: \"881f2dfd-9294-444d-9c05-18b72a9128a2\") " Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.715168 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "881f2dfd-9294-444d-9c05-18b72a9128a2" (UID: "881f2dfd-9294-444d-9c05-18b72a9128a2"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.715173 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/881f2dfd-9294-444d-9c05-18b72a9128a2-kube-api-access-r7b4n" (OuterVolumeSpecName: "kube-api-access-r7b4n") pod "881f2dfd-9294-444d-9c05-18b72a9128a2" (UID: "881f2dfd-9294-444d-9c05-18b72a9128a2"). InnerVolumeSpecName "kube-api-access-r7b4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.738165 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "881f2dfd-9294-444d-9c05-18b72a9128a2" (UID: "881f2dfd-9294-444d-9c05-18b72a9128a2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.759869 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-inventory" (OuterVolumeSpecName: "inventory") pod "881f2dfd-9294-444d-9c05-18b72a9128a2" (UID: "881f2dfd-9294-444d-9c05-18b72a9128a2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.809410 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7b4n\" (UniqueName: \"kubernetes.io/projected/881f2dfd-9294-444d-9c05-18b72a9128a2-kube-api-access-r7b4n\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.809449 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.809459 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:38 crc kubenswrapper[4960]: I1002 07:40:38.809469 4960 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/881f2dfd-9294-444d-9c05-18b72a9128a2-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.108703 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" event={"ID":"881f2dfd-9294-444d-9c05-18b72a9128a2","Type":"ContainerDied","Data":"5ee1145348db2c67a794b6c7dfcad767baecf87c9ee3459c9ad802efed130667"} Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.108790 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ee1145348db2c67a794b6c7dfcad767baecf87c9ee3459c9ad802efed130667" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.108813 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.217869 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b"] Oct 02 07:40:39 crc kubenswrapper[4960]: E1002 07:40:39.218428 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="extract-content" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.218445 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="extract-content" Oct 02 07:40:39 crc kubenswrapper[4960]: E1002 07:40:39.218462 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881f2dfd-9294-444d-9c05-18b72a9128a2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.218470 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="881f2dfd-9294-444d-9c05-18b72a9128a2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:40:39 crc kubenswrapper[4960]: E1002 07:40:39.218486 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="registry-server" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.218494 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="registry-server" Oct 02 07:40:39 crc kubenswrapper[4960]: E1002 07:40:39.230147 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="extract-utilities" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.230197 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="extract-utilities" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.230711 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb5e91eb-0f69-43fe-9b3d-9a75608c170a" containerName="registry-server" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.230739 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="881f2dfd-9294-444d-9c05-18b72a9128a2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.231464 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b"] Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.231567 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.234155 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.234480 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.234873 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.234956 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.320989 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.321054 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx8zz\" (UniqueName: \"kubernetes.io/projected/44e92992-9dec-4817-a070-035bde48c556-kube-api-access-zx8zz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.321255 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.422913 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.423122 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.423172 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx8zz\" (UniqueName: \"kubernetes.io/projected/44e92992-9dec-4817-a070-035bde48c556-kube-api-access-zx8zz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.428350 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.433276 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.444449 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx8zz\" (UniqueName: \"kubernetes.io/projected/44e92992-9dec-4817-a070-035bde48c556-kube-api-access-zx8zz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.578355 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:40:39 crc kubenswrapper[4960]: I1002 07:40:39.987465 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b"] Oct 02 07:40:40 crc kubenswrapper[4960]: I1002 07:40:40.119065 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" event={"ID":"44e92992-9dec-4817-a070-035bde48c556","Type":"ContainerStarted","Data":"fbac58552ffb65a5f82af3309c891a015b03bfa82d6aea19cc7bc05d0dbf0205"} Oct 02 07:40:41 crc kubenswrapper[4960]: I1002 07:40:41.131926 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" event={"ID":"44e92992-9dec-4817-a070-035bde48c556","Type":"ContainerStarted","Data":"f6b3d4d3ac0772db8208b0643fc13ec6b76610dfa4a9c4d64e5625c1843f9d40"} Oct 02 07:40:41 crc kubenswrapper[4960]: I1002 07:40:41.166512 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" podStartSLOduration=1.6523632 podStartE2EDuration="2.166487014s" podCreationTimestamp="2025-10-02 07:40:39 +0000 UTC" firstStartedPulling="2025-10-02 07:40:39.987124581 +0000 UTC m=+1461.019070868" lastFinishedPulling="2025-10-02 07:40:40.501248395 +0000 UTC m=+1461.533194682" observedRunningTime="2025-10-02 07:40:41.160529235 +0000 UTC m=+1462.192475522" watchObservedRunningTime="2025-10-02 07:40:41.166487014 +0000 UTC m=+1462.198433291" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.220660 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-45xbr"] Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.225479 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.242680 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-45xbr"] Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.396563 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-catalog-content\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.397166 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d85z8\" (UniqueName: \"kubernetes.io/projected/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-kube-api-access-d85z8\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.397412 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-utilities\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.499880 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d85z8\" (UniqueName: \"kubernetes.io/projected/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-kube-api-access-d85z8\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.500024 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-utilities\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.500088 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-catalog-content\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.501053 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-utilities\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.500969 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-catalog-content\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.523688 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d85z8\" (UniqueName: \"kubernetes.io/projected/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-kube-api-access-d85z8\") pod \"certified-operators-45xbr\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:42 crc kubenswrapper[4960]: I1002 07:40:42.553745 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:43 crc kubenswrapper[4960]: I1002 07:40:43.050463 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-45xbr"] Oct 02 07:40:43 crc kubenswrapper[4960]: I1002 07:40:43.155523 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45xbr" event={"ID":"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2","Type":"ContainerStarted","Data":"deee60e482a016778c01cf9d9a5c8e3f4d4f3b63c4ff463a3f1d7fcb9b74f35d"} Oct 02 07:40:44 crc kubenswrapper[4960]: I1002 07:40:44.171850 4960 generic.go:334] "Generic (PLEG): container finished" podID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerID="ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461" exitCode=0 Oct 02 07:40:44 crc kubenswrapper[4960]: I1002 07:40:44.171925 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45xbr" event={"ID":"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2","Type":"ContainerDied","Data":"ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461"} Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.218210 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45xbr" event={"ID":"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2","Type":"ContainerStarted","Data":"894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a"} Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.605266 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qtv55"] Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.608877 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.616160 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtv55"] Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.703211 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-utilities\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.703265 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfqc2\" (UniqueName: \"kubernetes.io/projected/e1db05ff-9dfc-4799-bb68-50c004b07811-kube-api-access-lfqc2\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.703548 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-catalog-content\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.805535 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-utilities\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.805595 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfqc2\" (UniqueName: \"kubernetes.io/projected/e1db05ff-9dfc-4799-bb68-50c004b07811-kube-api-access-lfqc2\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.805733 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-catalog-content\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.806097 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-utilities\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.806585 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-catalog-content\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.828587 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfqc2\" (UniqueName: \"kubernetes.io/projected/e1db05ff-9dfc-4799-bb68-50c004b07811-kube-api-access-lfqc2\") pod \"redhat-operators-qtv55\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:46 crc kubenswrapper[4960]: I1002 07:40:46.949891 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:47 crc kubenswrapper[4960]: I1002 07:40:47.235625 4960 generic.go:334] "Generic (PLEG): container finished" podID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerID="894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a" exitCode=0 Oct 02 07:40:47 crc kubenswrapper[4960]: I1002 07:40:47.235997 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45xbr" event={"ID":"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2","Type":"ContainerDied","Data":"894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a"} Oct 02 07:40:47 crc kubenswrapper[4960]: W1002 07:40:47.455512 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1db05ff_9dfc_4799_bb68_50c004b07811.slice/crio-3bba14c5148a36288315768f31bdd5e8f84fa89740c369d4fe01f82fbfa4b394 WatchSource:0}: Error finding container 3bba14c5148a36288315768f31bdd5e8f84fa89740c369d4fe01f82fbfa4b394: Status 404 returned error can't find the container with id 3bba14c5148a36288315768f31bdd5e8f84fa89740c369d4fe01f82fbfa4b394 Oct 02 07:40:47 crc kubenswrapper[4960]: I1002 07:40:47.462191 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtv55"] Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.249641 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45xbr" event={"ID":"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2","Type":"ContainerStarted","Data":"a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab"} Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.253579 4960 generic.go:334] "Generic (PLEG): container finished" podID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerID="0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa" exitCode=0 Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.253610 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtv55" event={"ID":"e1db05ff-9dfc-4799-bb68-50c004b07811","Type":"ContainerDied","Data":"0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa"} Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.253649 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtv55" event={"ID":"e1db05ff-9dfc-4799-bb68-50c004b07811","Type":"ContainerStarted","Data":"3bba14c5148a36288315768f31bdd5e8f84fa89740c369d4fe01f82fbfa4b394"} Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.280420 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-45xbr" podStartSLOduration=2.664001495 podStartE2EDuration="6.280400012s" podCreationTimestamp="2025-10-02 07:40:42 +0000 UTC" firstStartedPulling="2025-10-02 07:40:44.174792815 +0000 UTC m=+1465.206739132" lastFinishedPulling="2025-10-02 07:40:47.791191362 +0000 UTC m=+1468.823137649" observedRunningTime="2025-10-02 07:40:48.276569554 +0000 UTC m=+1469.308515861" watchObservedRunningTime="2025-10-02 07:40:48.280400012 +0000 UTC m=+1469.312346289" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.795114 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gbx2s"] Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.799622 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.839416 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbx2s"] Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.853205 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-catalog-content\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.853263 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-utilities\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.853319 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2mqx\" (UniqueName: \"kubernetes.io/projected/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-kube-api-access-p2mqx\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.956419 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-catalog-content\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.956481 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-utilities\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.956552 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2mqx\" (UniqueName: \"kubernetes.io/projected/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-kube-api-access-p2mqx\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.957695 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-utilities\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.957812 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-catalog-content\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:48 crc kubenswrapper[4960]: I1002 07:40:48.983178 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2mqx\" (UniqueName: \"kubernetes.io/projected/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-kube-api-access-p2mqx\") pod \"community-operators-gbx2s\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:49 crc kubenswrapper[4960]: I1002 07:40:49.141405 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:49 crc kubenswrapper[4960]: I1002 07:40:49.740550 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbx2s"] Oct 02 07:40:49 crc kubenswrapper[4960]: W1002 07:40:49.743704 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1586e2aa_73f6_4e06_bb1c_069a64ca3e61.slice/crio-8bbf378f0fb78bae1f9d46c38024b04b4c0d5ea14bb57c4547f9eab82f899ec4 WatchSource:0}: Error finding container 8bbf378f0fb78bae1f9d46c38024b04b4c0d5ea14bb57c4547f9eab82f899ec4: Status 404 returned error can't find the container with id 8bbf378f0fb78bae1f9d46c38024b04b4c0d5ea14bb57c4547f9eab82f899ec4 Oct 02 07:40:50 crc kubenswrapper[4960]: E1002 07:40:50.157479 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1db05ff_9dfc_4799_bb68_50c004b07811.slice/crio-ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1586e2aa_73f6_4e06_bb1c_069a64ca3e61.slice/crio-1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:40:50 crc kubenswrapper[4960]: I1002 07:40:50.286828 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerStarted","Data":"1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242"} Oct 02 07:40:50 crc kubenswrapper[4960]: I1002 07:40:50.286956 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerStarted","Data":"8bbf378f0fb78bae1f9d46c38024b04b4c0d5ea14bb57c4547f9eab82f899ec4"} Oct 02 07:40:50 crc kubenswrapper[4960]: I1002 07:40:50.289526 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtv55" event={"ID":"e1db05ff-9dfc-4799-bb68-50c004b07811","Type":"ContainerStarted","Data":"ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab"} Oct 02 07:40:51 crc kubenswrapper[4960]: I1002 07:40:51.317228 4960 generic.go:334] "Generic (PLEG): container finished" podID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerID="ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab" exitCode=0 Oct 02 07:40:51 crc kubenswrapper[4960]: I1002 07:40:51.318798 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtv55" event={"ID":"e1db05ff-9dfc-4799-bb68-50c004b07811","Type":"ContainerDied","Data":"ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab"} Oct 02 07:40:52 crc kubenswrapper[4960]: I1002 07:40:52.329583 4960 generic.go:334] "Generic (PLEG): container finished" podID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerID="1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242" exitCode=0 Oct 02 07:40:52 crc kubenswrapper[4960]: I1002 07:40:52.343406 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerDied","Data":"1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242"} Oct 02 07:40:52 crc kubenswrapper[4960]: I1002 07:40:52.343491 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtv55" event={"ID":"e1db05ff-9dfc-4799-bb68-50c004b07811","Type":"ContainerStarted","Data":"8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde"} Oct 02 07:40:52 crc kubenswrapper[4960]: I1002 07:40:52.362513 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qtv55" podStartSLOduration=2.770442194 podStartE2EDuration="6.362461263s" podCreationTimestamp="2025-10-02 07:40:46 +0000 UTC" firstStartedPulling="2025-10-02 07:40:48.256224919 +0000 UTC m=+1469.288171256" lastFinishedPulling="2025-10-02 07:40:51.848244028 +0000 UTC m=+1472.880190325" observedRunningTime="2025-10-02 07:40:52.353876461 +0000 UTC m=+1473.385822748" watchObservedRunningTime="2025-10-02 07:40:52.362461263 +0000 UTC m=+1473.394407550" Oct 02 07:40:52 crc kubenswrapper[4960]: I1002 07:40:52.554241 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:52 crc kubenswrapper[4960]: I1002 07:40:52.554718 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:52 crc kubenswrapper[4960]: I1002 07:40:52.621582 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:53 crc kubenswrapper[4960]: I1002 07:40:53.353268 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerStarted","Data":"bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a"} Oct 02 07:40:53 crc kubenswrapper[4960]: I1002 07:40:53.420097 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:55 crc kubenswrapper[4960]: I1002 07:40:55.183133 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-45xbr"] Oct 02 07:40:55 crc kubenswrapper[4960]: I1002 07:40:55.380650 4960 generic.go:334] "Generic (PLEG): container finished" podID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerID="bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a" exitCode=0 Oct 02 07:40:55 crc kubenswrapper[4960]: I1002 07:40:55.380720 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerDied","Data":"bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a"} Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.393010 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerStarted","Data":"1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3"} Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.393194 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-45xbr" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="registry-server" containerID="cri-o://a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab" gracePeriod=2 Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.425927 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gbx2s" podStartSLOduration=4.88803292 podStartE2EDuration="8.42590299s" podCreationTimestamp="2025-10-02 07:40:48 +0000 UTC" firstStartedPulling="2025-10-02 07:40:52.346358789 +0000 UTC m=+1473.378305076" lastFinishedPulling="2025-10-02 07:40:55.884228859 +0000 UTC m=+1476.916175146" observedRunningTime="2025-10-02 07:40:56.416433222 +0000 UTC m=+1477.448379509" watchObservedRunningTime="2025-10-02 07:40:56.42590299 +0000 UTC m=+1477.457849277" Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.910130 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.950416 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.950496 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.966753 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d85z8\" (UniqueName: \"kubernetes.io/projected/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-kube-api-access-d85z8\") pod \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.966920 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-utilities\") pod \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.967008 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-catalog-content\") pod \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\" (UID: \"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2\") " Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.967602 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-utilities" (OuterVolumeSpecName: "utilities") pod "e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" (UID: "e2cb5ba9-34cf-4d16-bbab-acff3a332ba2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:40:56 crc kubenswrapper[4960]: I1002 07:40:56.977012 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-kube-api-access-d85z8" (OuterVolumeSpecName: "kube-api-access-d85z8") pod "e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" (UID: "e2cb5ba9-34cf-4d16-bbab-acff3a332ba2"). InnerVolumeSpecName "kube-api-access-d85z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.028528 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" (UID: "e2cb5ba9-34cf-4d16-bbab-acff3a332ba2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.068468 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.068509 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d85z8\" (UniqueName: \"kubernetes.io/projected/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-kube-api-access-d85z8\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.068523 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.407529 4960 generic.go:334] "Generic (PLEG): container finished" podID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerID="a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab" exitCode=0 Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.407591 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45xbr" event={"ID":"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2","Type":"ContainerDied","Data":"a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab"} Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.407613 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45xbr" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.407638 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45xbr" event={"ID":"e2cb5ba9-34cf-4d16-bbab-acff3a332ba2","Type":"ContainerDied","Data":"deee60e482a016778c01cf9d9a5c8e3f4d4f3b63c4ff463a3f1d7fcb9b74f35d"} Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.407664 4960 scope.go:117] "RemoveContainer" containerID="a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.448044 4960 scope.go:117] "RemoveContainer" containerID="894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.452207 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-45xbr"] Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.466649 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-45xbr"] Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.500615 4960 scope.go:117] "RemoveContainer" containerID="ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.526518 4960 scope.go:117] "RemoveContainer" containerID="a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab" Oct 02 07:40:57 crc kubenswrapper[4960]: E1002 07:40:57.527212 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab\": container with ID starting with a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab not found: ID does not exist" containerID="a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.527250 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab"} err="failed to get container status \"a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab\": rpc error: code = NotFound desc = could not find container \"a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab\": container with ID starting with a92bb2fee80d227ce2e9e9625cbda32d99f5581620bd9343f7cb4c62bea0aeab not found: ID does not exist" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.527276 4960 scope.go:117] "RemoveContainer" containerID="894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a" Oct 02 07:40:57 crc kubenswrapper[4960]: E1002 07:40:57.527735 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a\": container with ID starting with 894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a not found: ID does not exist" containerID="894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.527761 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a"} err="failed to get container status \"894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a\": rpc error: code = NotFound desc = could not find container \"894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a\": container with ID starting with 894a85bf6a6be3187514a048f6a692a4cca6a9a22593467e6bdb1177b6b4bf0a not found: ID does not exist" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.527775 4960 scope.go:117] "RemoveContainer" containerID="ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461" Oct 02 07:40:57 crc kubenswrapper[4960]: E1002 07:40:57.528543 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461\": container with ID starting with ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461 not found: ID does not exist" containerID="ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461" Oct 02 07:40:57 crc kubenswrapper[4960]: I1002 07:40:57.528562 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461"} err="failed to get container status \"ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461\": rpc error: code = NotFound desc = could not find container \"ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461\": container with ID starting with ea76b4aa896ea6d56a53a1e9d7e47aa387cc9419fbfff0b0020421de8a551461 not found: ID does not exist" Oct 02 07:40:58 crc kubenswrapper[4960]: I1002 07:40:58.017245 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qtv55" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="registry-server" probeResult="failure" output=< Oct 02 07:40:58 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 07:40:58 crc kubenswrapper[4960]: > Oct 02 07:40:58 crc kubenswrapper[4960]: I1002 07:40:58.345494 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" path="/var/lib/kubelet/pods/e2cb5ba9-34cf-4d16-bbab-acff3a332ba2/volumes" Oct 02 07:40:59 crc kubenswrapper[4960]: I1002 07:40:59.142601 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:59 crc kubenswrapper[4960]: I1002 07:40:59.143065 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:40:59 crc kubenswrapper[4960]: I1002 07:40:59.149959 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:40:59 crc kubenswrapper[4960]: I1002 07:40:59.150043 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:40:59 crc kubenswrapper[4960]: I1002 07:40:59.239143 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:41:08 crc kubenswrapper[4960]: I1002 07:41:08.007150 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qtv55" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="registry-server" probeResult="failure" output=< Oct 02 07:41:08 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 07:41:08 crc kubenswrapper[4960]: > Oct 02 07:41:09 crc kubenswrapper[4960]: I1002 07:41:09.221730 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:41:09 crc kubenswrapper[4960]: I1002 07:41:09.285312 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbx2s"] Oct 02 07:41:09 crc kubenswrapper[4960]: I1002 07:41:09.541792 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gbx2s" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="registry-server" containerID="cri-o://1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3" gracePeriod=2 Oct 02 07:41:09 crc kubenswrapper[4960]: I1002 07:41:09.975665 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.066610 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2mqx\" (UniqueName: \"kubernetes.io/projected/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-kube-api-access-p2mqx\") pod \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.066715 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-catalog-content\") pod \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.066871 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-utilities\") pod \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\" (UID: \"1586e2aa-73f6-4e06-bb1c-069a64ca3e61\") " Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.068628 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-utilities" (OuterVolumeSpecName: "utilities") pod "1586e2aa-73f6-4e06-bb1c-069a64ca3e61" (UID: "1586e2aa-73f6-4e06-bb1c-069a64ca3e61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.073946 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-kube-api-access-p2mqx" (OuterVolumeSpecName: "kube-api-access-p2mqx") pod "1586e2aa-73f6-4e06-bb1c-069a64ca3e61" (UID: "1586e2aa-73f6-4e06-bb1c-069a64ca3e61"). InnerVolumeSpecName "kube-api-access-p2mqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.136929 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1586e2aa-73f6-4e06-bb1c-069a64ca3e61" (UID: "1586e2aa-73f6-4e06-bb1c-069a64ca3e61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.170534 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2mqx\" (UniqueName: \"kubernetes.io/projected/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-kube-api-access-p2mqx\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.170603 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.170617 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1586e2aa-73f6-4e06-bb1c-069a64ca3e61-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.554963 4960 generic.go:334] "Generic (PLEG): container finished" podID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerID="1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3" exitCode=0 Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.555052 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerDied","Data":"1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3"} Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.555092 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbx2s" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.555134 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbx2s" event={"ID":"1586e2aa-73f6-4e06-bb1c-069a64ca3e61","Type":"ContainerDied","Data":"8bbf378f0fb78bae1f9d46c38024b04b4c0d5ea14bb57c4547f9eab82f899ec4"} Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.555162 4960 scope.go:117] "RemoveContainer" containerID="1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.586014 4960 scope.go:117] "RemoveContainer" containerID="bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.588086 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbx2s"] Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.601010 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gbx2s"] Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.612583 4960 scope.go:117] "RemoveContainer" containerID="1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.661771 4960 scope.go:117] "RemoveContainer" containerID="1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3" Oct 02 07:41:10 crc kubenswrapper[4960]: E1002 07:41:10.662378 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3\": container with ID starting with 1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3 not found: ID does not exist" containerID="1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.662423 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3"} err="failed to get container status \"1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3\": rpc error: code = NotFound desc = could not find container \"1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3\": container with ID starting with 1a24a34563728c5bb67e0f0e56505058c0f30885b1eed1f91068736dcb10e2a3 not found: ID does not exist" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.662449 4960 scope.go:117] "RemoveContainer" containerID="bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a" Oct 02 07:41:10 crc kubenswrapper[4960]: E1002 07:41:10.662864 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a\": container with ID starting with bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a not found: ID does not exist" containerID="bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.662900 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a"} err="failed to get container status \"bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a\": rpc error: code = NotFound desc = could not find container \"bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a\": container with ID starting with bf9922518eec72357c76c3aa0e27876119336cebc8f61dd927c8db9bbd9b361a not found: ID does not exist" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.662928 4960 scope.go:117] "RemoveContainer" containerID="1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242" Oct 02 07:41:10 crc kubenswrapper[4960]: E1002 07:41:10.663311 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242\": container with ID starting with 1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242 not found: ID does not exist" containerID="1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242" Oct 02 07:41:10 crc kubenswrapper[4960]: I1002 07:41:10.663366 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242"} err="failed to get container status \"1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242\": rpc error: code = NotFound desc = could not find container \"1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242\": container with ID starting with 1cc023a293669b663855948ebc0a0b5b792618bbc7dd4f3e5d07f0c499789242 not found: ID does not exist" Oct 02 07:41:12 crc kubenswrapper[4960]: I1002 07:41:12.340069 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" path="/var/lib/kubelet/pods/1586e2aa-73f6-4e06-bb1c-069a64ca3e61/volumes" Oct 02 07:41:17 crc kubenswrapper[4960]: I1002 07:41:17.008158 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:41:17 crc kubenswrapper[4960]: I1002 07:41:17.074356 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:41:17 crc kubenswrapper[4960]: I1002 07:41:17.805371 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qtv55"] Oct 02 07:41:18 crc kubenswrapper[4960]: I1002 07:41:18.652520 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qtv55" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="registry-server" containerID="cri-o://8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde" gracePeriod=2 Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.122000 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.281943 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-catalog-content\") pod \"e1db05ff-9dfc-4799-bb68-50c004b07811\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.282076 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfqc2\" (UniqueName: \"kubernetes.io/projected/e1db05ff-9dfc-4799-bb68-50c004b07811-kube-api-access-lfqc2\") pod \"e1db05ff-9dfc-4799-bb68-50c004b07811\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.283125 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-utilities\") pod \"e1db05ff-9dfc-4799-bb68-50c004b07811\" (UID: \"e1db05ff-9dfc-4799-bb68-50c004b07811\") " Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.284034 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-utilities" (OuterVolumeSpecName: "utilities") pod "e1db05ff-9dfc-4799-bb68-50c004b07811" (UID: "e1db05ff-9dfc-4799-bb68-50c004b07811"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.291207 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1db05ff-9dfc-4799-bb68-50c004b07811-kube-api-access-lfqc2" (OuterVolumeSpecName: "kube-api-access-lfqc2") pod "e1db05ff-9dfc-4799-bb68-50c004b07811" (UID: "e1db05ff-9dfc-4799-bb68-50c004b07811"). InnerVolumeSpecName "kube-api-access-lfqc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.384795 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfqc2\" (UniqueName: \"kubernetes.io/projected/e1db05ff-9dfc-4799-bb68-50c004b07811-kube-api-access-lfqc2\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.384827 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.393970 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1db05ff-9dfc-4799-bb68-50c004b07811" (UID: "e1db05ff-9dfc-4799-bb68-50c004b07811"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.487064 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1db05ff-9dfc-4799-bb68-50c004b07811-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.665616 4960 generic.go:334] "Generic (PLEG): container finished" podID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerID="8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde" exitCode=0 Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.665712 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtv55" event={"ID":"e1db05ff-9dfc-4799-bb68-50c004b07811","Type":"ContainerDied","Data":"8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde"} Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.665797 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtv55" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.666846 4960 scope.go:117] "RemoveContainer" containerID="8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.666743 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtv55" event={"ID":"e1db05ff-9dfc-4799-bb68-50c004b07811","Type":"ContainerDied","Data":"3bba14c5148a36288315768f31bdd5e8f84fa89740c369d4fe01f82fbfa4b394"} Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.703548 4960 scope.go:117] "RemoveContainer" containerID="ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.731341 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qtv55"] Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.740903 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qtv55"] Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.745353 4960 scope.go:117] "RemoveContainer" containerID="0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.778512 4960 scope.go:117] "RemoveContainer" containerID="8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde" Oct 02 07:41:19 crc kubenswrapper[4960]: E1002 07:41:19.779144 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde\": container with ID starting with 8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde not found: ID does not exist" containerID="8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.779194 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde"} err="failed to get container status \"8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde\": rpc error: code = NotFound desc = could not find container \"8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde\": container with ID starting with 8992a2aa5e5134465c5e6044476be4eafc4133a5a034cffd2b930d592759dcde not found: ID does not exist" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.779227 4960 scope.go:117] "RemoveContainer" containerID="ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab" Oct 02 07:41:19 crc kubenswrapper[4960]: E1002 07:41:19.779808 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab\": container with ID starting with ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab not found: ID does not exist" containerID="ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.779859 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab"} err="failed to get container status \"ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab\": rpc error: code = NotFound desc = could not find container \"ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab\": container with ID starting with ef912e30bcb7a2c50e13a9f957dbcc5f4efe575b2bf275179bff731c9c59beab not found: ID does not exist" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.779895 4960 scope.go:117] "RemoveContainer" containerID="0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa" Oct 02 07:41:19 crc kubenswrapper[4960]: E1002 07:41:19.780504 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa\": container with ID starting with 0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa not found: ID does not exist" containerID="0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa" Oct 02 07:41:19 crc kubenswrapper[4960]: I1002 07:41:19.780560 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa"} err="failed to get container status \"0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa\": rpc error: code = NotFound desc = could not find container \"0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa\": container with ID starting with 0fb0ec89567fc530e401889d1c04ea709b92e1b38f2dd6fa61fa7fc888dfadaa not found: ID does not exist" Oct 02 07:41:20 crc kubenswrapper[4960]: I1002 07:41:20.352426 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" path="/var/lib/kubelet/pods/e1db05ff-9dfc-4799-bb68-50c004b07811/volumes" Oct 02 07:41:29 crc kubenswrapper[4960]: I1002 07:41:29.150122 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:41:29 crc kubenswrapper[4960]: I1002 07:41:29.151123 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:41:53 crc kubenswrapper[4960]: I1002 07:41:53.061850 4960 generic.go:334] "Generic (PLEG): container finished" podID="44e92992-9dec-4817-a070-035bde48c556" containerID="f6b3d4d3ac0772db8208b0643fc13ec6b76610dfa4a9c4d64e5625c1843f9d40" exitCode=0 Oct 02 07:41:53 crc kubenswrapper[4960]: I1002 07:41:53.061953 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" event={"ID":"44e92992-9dec-4817-a070-035bde48c556","Type":"ContainerDied","Data":"f6b3d4d3ac0772db8208b0643fc13ec6b76610dfa4a9c4d64e5625c1843f9d40"} Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.541723 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.623895 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx8zz\" (UniqueName: \"kubernetes.io/projected/44e92992-9dec-4817-a070-035bde48c556-kube-api-access-zx8zz\") pod \"44e92992-9dec-4817-a070-035bde48c556\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.624295 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-ssh-key\") pod \"44e92992-9dec-4817-a070-035bde48c556\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.624377 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-inventory\") pod \"44e92992-9dec-4817-a070-035bde48c556\" (UID: \"44e92992-9dec-4817-a070-035bde48c556\") " Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.635837 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44e92992-9dec-4817-a070-035bde48c556-kube-api-access-zx8zz" (OuterVolumeSpecName: "kube-api-access-zx8zz") pod "44e92992-9dec-4817-a070-035bde48c556" (UID: "44e92992-9dec-4817-a070-035bde48c556"). InnerVolumeSpecName "kube-api-access-zx8zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.656072 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-inventory" (OuterVolumeSpecName: "inventory") pod "44e92992-9dec-4817-a070-035bde48c556" (UID: "44e92992-9dec-4817-a070-035bde48c556"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.666215 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "44e92992-9dec-4817-a070-035bde48c556" (UID: "44e92992-9dec-4817-a070-035bde48c556"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.727719 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.727761 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e92992-9dec-4817-a070-035bde48c556-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:54 crc kubenswrapper[4960]: I1002 07:41:54.727776 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx8zz\" (UniqueName: \"kubernetes.io/projected/44e92992-9dec-4817-a070-035bde48c556-kube-api-access-zx8zz\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.095733 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" event={"ID":"44e92992-9dec-4817-a070-035bde48c556","Type":"ContainerDied","Data":"fbac58552ffb65a5f82af3309c891a015b03bfa82d6aea19cc7bc05d0dbf0205"} Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.096423 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbac58552ffb65a5f82af3309c891a015b03bfa82d6aea19cc7bc05d0dbf0205" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.095873 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.206619 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x"] Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.207449 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="extract-utilities" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.207548 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="extract-utilities" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.207642 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="extract-utilities" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.207768 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="extract-utilities" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.207860 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="extract-content" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.207950 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="extract-content" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.208052 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.208132 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.208219 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.208285 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.208366 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="extract-content" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.208438 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="extract-content" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.208520 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.208592 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.208660 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="extract-content" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.208723 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="extract-content" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.208797 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44e92992-9dec-4817-a070-035bde48c556" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.208872 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="44e92992-9dec-4817-a070-035bde48c556" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:41:55 crc kubenswrapper[4960]: E1002 07:41:55.208945 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="extract-utilities" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.209042 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="extract-utilities" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.209398 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="1586e2aa-73f6-4e06-bb1c-069a64ca3e61" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.209517 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2cb5ba9-34cf-4d16-bbab-acff3a332ba2" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.209597 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="44e92992-9dec-4817-a070-035bde48c556" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.209792 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1db05ff-9dfc-4799-bb68-50c004b07811" containerName="registry-server" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.210786 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.214480 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.215893 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.216298 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.219126 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.232009 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x"] Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.346185 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.346565 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.346760 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vz7k\" (UniqueName: \"kubernetes.io/projected/f9302552-a6d8-43f6-93d7-05db5a3a9e30-kube-api-access-5vz7k\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.448206 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.448340 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vz7k\" (UniqueName: \"kubernetes.io/projected/f9302552-a6d8-43f6-93d7-05db5a3a9e30-kube-api-access-5vz7k\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.448599 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.454786 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.466994 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vz7k\" (UniqueName: \"kubernetes.io/projected/f9302552-a6d8-43f6-93d7-05db5a3a9e30-kube-api-access-5vz7k\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.471635 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:55 crc kubenswrapper[4960]: I1002 07:41:55.546577 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:41:56 crc kubenswrapper[4960]: I1002 07:41:56.141265 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x"] Oct 02 07:41:57 crc kubenswrapper[4960]: I1002 07:41:57.115956 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" event={"ID":"f9302552-a6d8-43f6-93d7-05db5a3a9e30","Type":"ContainerStarted","Data":"72151a5c2fc6d9a50f9f5eb1970d291b79db641639cb3d48ff61df9d042eb246"} Oct 02 07:41:57 crc kubenswrapper[4960]: I1002 07:41:57.117797 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" event={"ID":"f9302552-a6d8-43f6-93d7-05db5a3a9e30","Type":"ContainerStarted","Data":"32f2b8c782c1b56876816d41c7bce07b80882f216009d69b34a5c19f7b9df71a"} Oct 02 07:41:57 crc kubenswrapper[4960]: I1002 07:41:57.143168 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" podStartSLOduration=1.6196827310000002 podStartE2EDuration="2.143145182s" podCreationTimestamp="2025-10-02 07:41:55 +0000 UTC" firstStartedPulling="2025-10-02 07:41:56.152363947 +0000 UTC m=+1537.184310234" lastFinishedPulling="2025-10-02 07:41:56.675826408 +0000 UTC m=+1537.707772685" observedRunningTime="2025-10-02 07:41:57.136247686 +0000 UTC m=+1538.168193973" watchObservedRunningTime="2025-10-02 07:41:57.143145182 +0000 UTC m=+1538.175091469" Oct 02 07:41:59 crc kubenswrapper[4960]: I1002 07:41:59.150150 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:41:59 crc kubenswrapper[4960]: I1002 07:41:59.150704 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:41:59 crc kubenswrapper[4960]: I1002 07:41:59.150785 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:41:59 crc kubenswrapper[4960]: I1002 07:41:59.151570 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:41:59 crc kubenswrapper[4960]: I1002 07:41:59.151655 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" gracePeriod=600 Oct 02 07:41:59 crc kubenswrapper[4960]: E1002 07:41:59.280990 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:42:00 crc kubenswrapper[4960]: I1002 07:42:00.157140 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" exitCode=0 Oct 02 07:42:00 crc kubenswrapper[4960]: I1002 07:42:00.157238 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678"} Oct 02 07:42:00 crc kubenswrapper[4960]: I1002 07:42:00.158208 4960 scope.go:117] "RemoveContainer" containerID="f5bb746b73018c3c6e2b85b664f4b1896310fbe4f05d3269e3eee3a9625ffe9d" Oct 02 07:42:00 crc kubenswrapper[4960]: I1002 07:42:00.159402 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:42:00 crc kubenswrapper[4960]: E1002 07:42:00.159946 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:42:03 crc kubenswrapper[4960]: I1002 07:42:03.203557 4960 generic.go:334] "Generic (PLEG): container finished" podID="f9302552-a6d8-43f6-93d7-05db5a3a9e30" containerID="72151a5c2fc6d9a50f9f5eb1970d291b79db641639cb3d48ff61df9d042eb246" exitCode=0 Oct 02 07:42:03 crc kubenswrapper[4960]: I1002 07:42:03.203613 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" event={"ID":"f9302552-a6d8-43f6-93d7-05db5a3a9e30","Type":"ContainerDied","Data":"72151a5c2fc6d9a50f9f5eb1970d291b79db641639cb3d48ff61df9d042eb246"} Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.701318 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.790790 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vz7k\" (UniqueName: \"kubernetes.io/projected/f9302552-a6d8-43f6-93d7-05db5a3a9e30-kube-api-access-5vz7k\") pod \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.790951 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-ssh-key\") pod \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.791349 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-inventory\") pod \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\" (UID: \"f9302552-a6d8-43f6-93d7-05db5a3a9e30\") " Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.799305 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9302552-a6d8-43f6-93d7-05db5a3a9e30-kube-api-access-5vz7k" (OuterVolumeSpecName: "kube-api-access-5vz7k") pod "f9302552-a6d8-43f6-93d7-05db5a3a9e30" (UID: "f9302552-a6d8-43f6-93d7-05db5a3a9e30"). InnerVolumeSpecName "kube-api-access-5vz7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.821218 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-inventory" (OuterVolumeSpecName: "inventory") pod "f9302552-a6d8-43f6-93d7-05db5a3a9e30" (UID: "f9302552-a6d8-43f6-93d7-05db5a3a9e30"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.822874 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9302552-a6d8-43f6-93d7-05db5a3a9e30" (UID: "f9302552-a6d8-43f6-93d7-05db5a3a9e30"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.894192 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vz7k\" (UniqueName: \"kubernetes.io/projected/f9302552-a6d8-43f6-93d7-05db5a3a9e30-kube-api-access-5vz7k\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.894240 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:04 crc kubenswrapper[4960]: I1002 07:42:04.894254 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9302552-a6d8-43f6-93d7-05db5a3a9e30-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.232310 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" event={"ID":"f9302552-a6d8-43f6-93d7-05db5a3a9e30","Type":"ContainerDied","Data":"32f2b8c782c1b56876816d41c7bce07b80882f216009d69b34a5c19f7b9df71a"} Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.232366 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32f2b8c782c1b56876816d41c7bce07b80882f216009d69b34a5c19f7b9df71a" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.232406 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.338189 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl"] Oct 02 07:42:05 crc kubenswrapper[4960]: E1002 07:42:05.338710 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9302552-a6d8-43f6-93d7-05db5a3a9e30" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.338734 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9302552-a6d8-43f6-93d7-05db5a3a9e30" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.339185 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9302552-a6d8-43f6-93d7-05db5a3a9e30" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.340403 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.342828 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.343133 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.343695 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.345140 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.360343 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl"] Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.407883 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lljxq\" (UniqueName: \"kubernetes.io/projected/4077b2fa-ed35-41af-bca4-562ff6d9c100-kube-api-access-lljxq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.408099 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.408164 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.510886 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lljxq\" (UniqueName: \"kubernetes.io/projected/4077b2fa-ed35-41af-bca4-562ff6d9c100-kube-api-access-lljxq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.511017 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.511125 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.519940 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.520223 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.544674 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lljxq\" (UniqueName: \"kubernetes.io/projected/4077b2fa-ed35-41af-bca4-562ff6d9c100-kube-api-access-lljxq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8s4kl\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:05 crc kubenswrapper[4960]: I1002 07:42:05.676570 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:06 crc kubenswrapper[4960]: I1002 07:42:06.264224 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl"] Oct 02 07:42:06 crc kubenswrapper[4960]: I1002 07:42:06.274271 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:42:07 crc kubenswrapper[4960]: I1002 07:42:07.258058 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" event={"ID":"4077b2fa-ed35-41af-bca4-562ff6d9c100","Type":"ContainerStarted","Data":"58b10688097179490c642ef7ff2e726b40b19eeca457354427a375a4536d0392"} Oct 02 07:42:07 crc kubenswrapper[4960]: I1002 07:42:07.258654 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" event={"ID":"4077b2fa-ed35-41af-bca4-562ff6d9c100","Type":"ContainerStarted","Data":"cef0eff45020a4668076215d4b6d5138910539c951abe89a7fe27815eece68d0"} Oct 02 07:42:07 crc kubenswrapper[4960]: I1002 07:42:07.285148 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" podStartSLOduration=1.6365503810000002 podStartE2EDuration="2.285118153s" podCreationTimestamp="2025-10-02 07:42:05 +0000 UTC" firstStartedPulling="2025-10-02 07:42:06.274007152 +0000 UTC m=+1547.305953439" lastFinishedPulling="2025-10-02 07:42:06.922574924 +0000 UTC m=+1547.954521211" observedRunningTime="2025-10-02 07:42:07.275144235 +0000 UTC m=+1548.307090522" watchObservedRunningTime="2025-10-02 07:42:07.285118153 +0000 UTC m=+1548.317064460" Oct 02 07:42:09 crc kubenswrapper[4960]: I1002 07:42:09.058806 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-sck8h"] Oct 02 07:42:09 crc kubenswrapper[4960]: I1002 07:42:09.067741 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-sck8h"] Oct 02 07:42:10 crc kubenswrapper[4960]: I1002 07:42:10.349772 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a808721-6be5-4339-b4cc-770c6f288564" path="/var/lib/kubelet/pods/2a808721-6be5-4339-b4cc-770c6f288564/volumes" Oct 02 07:42:12 crc kubenswrapper[4960]: I1002 07:42:12.071792 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-xr5gr"] Oct 02 07:42:12 crc kubenswrapper[4960]: I1002 07:42:12.089666 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-xr5gr"] Oct 02 07:42:12 crc kubenswrapper[4960]: I1002 07:42:12.357329 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b98a8729-a22e-49e4-ae15-6ff17fb6c4b7" path="/var/lib/kubelet/pods/b98a8729-a22e-49e4-ae15-6ff17fb6c4b7/volumes" Oct 02 07:42:13 crc kubenswrapper[4960]: I1002 07:42:13.042531 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-p2r6q"] Oct 02 07:42:13 crc kubenswrapper[4960]: I1002 07:42:13.053949 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-p2r6q"] Oct 02 07:42:13 crc kubenswrapper[4960]: I1002 07:42:13.330619 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:42:13 crc kubenswrapper[4960]: E1002 07:42:13.330943 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:42:14 crc kubenswrapper[4960]: I1002 07:42:14.345133 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="109e5c0f-e848-4762-bdc8-1ae57e4f7fd7" path="/var/lib/kubelet/pods/109e5c0f-e848-4762-bdc8-1ae57e4f7fd7/volumes" Oct 02 07:42:19 crc kubenswrapper[4960]: I1002 07:42:19.043056 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ef5b-account-create-sdwnj"] Oct 02 07:42:19 crc kubenswrapper[4960]: I1002 07:42:19.053447 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ef5b-account-create-sdwnj"] Oct 02 07:42:20 crc kubenswrapper[4960]: I1002 07:42:20.352712 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44887da1-2bcd-482d-a53e-75e1ed397f84" path="/var/lib/kubelet/pods/44887da1-2bcd-482d-a53e-75e1ed397f84/volumes" Oct 02 07:42:24 crc kubenswrapper[4960]: I1002 07:42:24.034441 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8399-account-create-r2ck2"] Oct 02 07:42:24 crc kubenswrapper[4960]: I1002 07:42:24.046873 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8808-account-create-lk8m6"] Oct 02 07:42:24 crc kubenswrapper[4960]: I1002 07:42:24.059887 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8399-account-create-r2ck2"] Oct 02 07:42:24 crc kubenswrapper[4960]: I1002 07:42:24.070431 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8808-account-create-lk8m6"] Oct 02 07:42:24 crc kubenswrapper[4960]: I1002 07:42:24.368533 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fb0af53-ac0a-43d0-a651-49f1970f13f7" path="/var/lib/kubelet/pods/3fb0af53-ac0a-43d0-a651-49f1970f13f7/volumes" Oct 02 07:42:24 crc kubenswrapper[4960]: I1002 07:42:24.369624 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d33dff64-7c3f-404c-9eb4-5793c6bec19b" path="/var/lib/kubelet/pods/d33dff64-7c3f-404c-9eb4-5793c6bec19b/volumes" Oct 02 07:42:28 crc kubenswrapper[4960]: I1002 07:42:28.330460 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:42:28 crc kubenswrapper[4960]: E1002 07:42:28.331255 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:42:41 crc kubenswrapper[4960]: I1002 07:42:41.870923 4960 scope.go:117] "RemoveContainer" containerID="7416f347d2e27dbb509a6ad2ba9d46c21e6f28f8898f31098f0f2d07ee5ce2c2" Oct 02 07:42:41 crc kubenswrapper[4960]: I1002 07:42:41.932714 4960 scope.go:117] "RemoveContainer" containerID="7cb50b5d882242ae41e7d94e452a1d1fad0925228252dcea8f75ed4a513110c9" Oct 02 07:42:41 crc kubenswrapper[4960]: I1002 07:42:41.974027 4960 scope.go:117] "RemoveContainer" containerID="3226f9377160d51ee6cfdab42ebdfceeee9c1d29f98923450fcf0505b25cac57" Oct 02 07:42:42 crc kubenswrapper[4960]: I1002 07:42:42.018132 4960 scope.go:117] "RemoveContainer" containerID="c55ad313c401ecf9edf935b96cfc608b1cd48b63bc87ea5427af8d43c771cfe9" Oct 02 07:42:42 crc kubenswrapper[4960]: I1002 07:42:42.072453 4960 scope.go:117] "RemoveContainer" containerID="aacf31b34dfb947084057bc6b1d6e3a965ea5171f0bc592fb2a4fc44ff5dc802" Oct 02 07:42:42 crc kubenswrapper[4960]: I1002 07:42:42.104374 4960 scope.go:117] "RemoveContainer" containerID="86a0aeb86d65165f2bdaa438b63c62cc70b9f17c07ae8704e79d8c16d1ec0560" Oct 02 07:42:43 crc kubenswrapper[4960]: I1002 07:42:43.329944 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:42:43 crc kubenswrapper[4960]: E1002 07:42:43.330790 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.070838 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-lnmn2"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.081787 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-mfsh5"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.088563 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-w875j"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.096048 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-2kx2k"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.110140 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-mfsh5"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.117157 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-w875j"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.124210 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-lnmn2"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.133786 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-2kx2k"] Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.342446 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df3811d-f779-4df9-a3cc-6a74d7dc00b3" path="/var/lib/kubelet/pods/7df3811d-f779-4df9-a3cc-6a74d7dc00b3/volumes" Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.343308 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85344e88-9937-4cb2-903a-d62a423ce703" path="/var/lib/kubelet/pods/85344e88-9937-4cb2-903a-d62a423ce703/volumes" Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.344255 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15" path="/var/lib/kubelet/pods/ce1e9cd1-c5b1-4c9d-9f92-41e1ca3b2e15/volumes" Oct 02 07:42:46 crc kubenswrapper[4960]: I1002 07:42:46.345332 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0aea82a-faa5-46de-a545-54b2ac12c1bd" path="/var/lib/kubelet/pods/f0aea82a-faa5-46de-a545-54b2ac12c1bd/volumes" Oct 02 07:42:48 crc kubenswrapper[4960]: I1002 07:42:48.755350 4960 generic.go:334] "Generic (PLEG): container finished" podID="4077b2fa-ed35-41af-bca4-562ff6d9c100" containerID="58b10688097179490c642ef7ff2e726b40b19eeca457354427a375a4536d0392" exitCode=0 Oct 02 07:42:48 crc kubenswrapper[4960]: I1002 07:42:48.755479 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" event={"ID":"4077b2fa-ed35-41af-bca4-562ff6d9c100","Type":"ContainerDied","Data":"58b10688097179490c642ef7ff2e726b40b19eeca457354427a375a4536d0392"} Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.263158 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.380341 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-ssh-key\") pod \"4077b2fa-ed35-41af-bca4-562ff6d9c100\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.380599 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-inventory\") pod \"4077b2fa-ed35-41af-bca4-562ff6d9c100\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.380681 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lljxq\" (UniqueName: \"kubernetes.io/projected/4077b2fa-ed35-41af-bca4-562ff6d9c100-kube-api-access-lljxq\") pod \"4077b2fa-ed35-41af-bca4-562ff6d9c100\" (UID: \"4077b2fa-ed35-41af-bca4-562ff6d9c100\") " Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.391291 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4077b2fa-ed35-41af-bca4-562ff6d9c100-kube-api-access-lljxq" (OuterVolumeSpecName: "kube-api-access-lljxq") pod "4077b2fa-ed35-41af-bca4-562ff6d9c100" (UID: "4077b2fa-ed35-41af-bca4-562ff6d9c100"). InnerVolumeSpecName "kube-api-access-lljxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.417743 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4077b2fa-ed35-41af-bca4-562ff6d9c100" (UID: "4077b2fa-ed35-41af-bca4-562ff6d9c100"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.430267 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-inventory" (OuterVolumeSpecName: "inventory") pod "4077b2fa-ed35-41af-bca4-562ff6d9c100" (UID: "4077b2fa-ed35-41af-bca4-562ff6d9c100"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.484650 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.484716 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4077b2fa-ed35-41af-bca4-562ff6d9c100-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.484739 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lljxq\" (UniqueName: \"kubernetes.io/projected/4077b2fa-ed35-41af-bca4-562ff6d9c100-kube-api-access-lljxq\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.780559 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" event={"ID":"4077b2fa-ed35-41af-bca4-562ff6d9c100","Type":"ContainerDied","Data":"cef0eff45020a4668076215d4b6d5138910539c951abe89a7fe27815eece68d0"} Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.780610 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cef0eff45020a4668076215d4b6d5138910539c951abe89a7fe27815eece68d0" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.780677 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.884959 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92"] Oct 02 07:42:50 crc kubenswrapper[4960]: E1002 07:42:50.885373 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4077b2fa-ed35-41af-bca4-562ff6d9c100" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.885393 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4077b2fa-ed35-41af-bca4-562ff6d9c100" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.885596 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4077b2fa-ed35-41af-bca4-562ff6d9c100" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.886301 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.888775 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.888803 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.889026 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.889200 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.919280 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92"] Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.993915 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.994127 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhtfh\" (UniqueName: \"kubernetes.io/projected/4fccc04f-086d-4956-b703-9a7615989363-kube-api-access-jhtfh\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:50 crc kubenswrapper[4960]: I1002 07:42:50.994235 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.042521 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bcvcn"] Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.055426 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bcvcn"] Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.097066 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.097158 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhtfh\" (UniqueName: \"kubernetes.io/projected/4fccc04f-086d-4956-b703-9a7615989363-kube-api-access-jhtfh\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.097238 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.103669 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.112507 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.116516 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhtfh\" (UniqueName: \"kubernetes.io/projected/4fccc04f-086d-4956-b703-9a7615989363-kube-api-access-jhtfh\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.207683 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:51 crc kubenswrapper[4960]: I1002 07:42:51.778045 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92"] Oct 02 07:42:51 crc kubenswrapper[4960]: W1002 07:42:51.786630 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fccc04f_086d_4956_b703_9a7615989363.slice/crio-1b62b30c0be37c6ccb560b9bbaa88ef40448bbfbd0f9378aae1151e2d7ae1f40 WatchSource:0}: Error finding container 1b62b30c0be37c6ccb560b9bbaa88ef40448bbfbd0f9378aae1151e2d7ae1f40: Status 404 returned error can't find the container with id 1b62b30c0be37c6ccb560b9bbaa88ef40448bbfbd0f9378aae1151e2d7ae1f40 Oct 02 07:42:52 crc kubenswrapper[4960]: I1002 07:42:52.353187 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39377f09-4342-42bc-87a5-17639bdd8f99" path="/var/lib/kubelet/pods/39377f09-4342-42bc-87a5-17639bdd8f99/volumes" Oct 02 07:42:52 crc kubenswrapper[4960]: I1002 07:42:52.800672 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" event={"ID":"4fccc04f-086d-4956-b703-9a7615989363","Type":"ContainerStarted","Data":"7f3e3bf87367fa4cc7d2bb7d03f02691e5bb4878a2954f8c4844e82397b7143e"} Oct 02 07:42:52 crc kubenswrapper[4960]: I1002 07:42:52.800734 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" event={"ID":"4fccc04f-086d-4956-b703-9a7615989363","Type":"ContainerStarted","Data":"1b62b30c0be37c6ccb560b9bbaa88ef40448bbfbd0f9378aae1151e2d7ae1f40"} Oct 02 07:42:52 crc kubenswrapper[4960]: I1002 07:42:52.824718 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" podStartSLOduration=2.364147781 podStartE2EDuration="2.824691066s" podCreationTimestamp="2025-10-02 07:42:50 +0000 UTC" firstStartedPulling="2025-10-02 07:42:51.789539137 +0000 UTC m=+1592.821485424" lastFinishedPulling="2025-10-02 07:42:52.250082422 +0000 UTC m=+1593.282028709" observedRunningTime="2025-10-02 07:42:52.81679335 +0000 UTC m=+1593.848739647" watchObservedRunningTime="2025-10-02 07:42:52.824691066 +0000 UTC m=+1593.856637373" Oct 02 07:42:57 crc kubenswrapper[4960]: I1002 07:42:57.331196 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:42:57 crc kubenswrapper[4960]: E1002 07:42:57.331760 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:42:57 crc kubenswrapper[4960]: I1002 07:42:57.857161 4960 generic.go:334] "Generic (PLEG): container finished" podID="4fccc04f-086d-4956-b703-9a7615989363" containerID="7f3e3bf87367fa4cc7d2bb7d03f02691e5bb4878a2954f8c4844e82397b7143e" exitCode=0 Oct 02 07:42:57 crc kubenswrapper[4960]: I1002 07:42:57.857226 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" event={"ID":"4fccc04f-086d-4956-b703-9a7615989363","Type":"ContainerDied","Data":"7f3e3bf87367fa4cc7d2bb7d03f02691e5bb4878a2954f8c4844e82397b7143e"} Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.407506 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.487604 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhtfh\" (UniqueName: \"kubernetes.io/projected/4fccc04f-086d-4956-b703-9a7615989363-kube-api-access-jhtfh\") pod \"4fccc04f-086d-4956-b703-9a7615989363\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.487706 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-ssh-key\") pod \"4fccc04f-086d-4956-b703-9a7615989363\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.487752 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-inventory\") pod \"4fccc04f-086d-4956-b703-9a7615989363\" (UID: \"4fccc04f-086d-4956-b703-9a7615989363\") " Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.497099 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fccc04f-086d-4956-b703-9a7615989363-kube-api-access-jhtfh" (OuterVolumeSpecName: "kube-api-access-jhtfh") pod "4fccc04f-086d-4956-b703-9a7615989363" (UID: "4fccc04f-086d-4956-b703-9a7615989363"). InnerVolumeSpecName "kube-api-access-jhtfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.525885 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-inventory" (OuterVolumeSpecName: "inventory") pod "4fccc04f-086d-4956-b703-9a7615989363" (UID: "4fccc04f-086d-4956-b703-9a7615989363"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.526368 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4fccc04f-086d-4956-b703-9a7615989363" (UID: "4fccc04f-086d-4956-b703-9a7615989363"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.591415 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhtfh\" (UniqueName: \"kubernetes.io/projected/4fccc04f-086d-4956-b703-9a7615989363-kube-api-access-jhtfh\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.591471 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.591483 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fccc04f-086d-4956-b703-9a7615989363-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.882185 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" event={"ID":"4fccc04f-086d-4956-b703-9a7615989363","Type":"ContainerDied","Data":"1b62b30c0be37c6ccb560b9bbaa88ef40448bbfbd0f9378aae1151e2d7ae1f40"} Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.882245 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b62b30c0be37c6ccb560b9bbaa88ef40448bbfbd0f9378aae1151e2d7ae1f40" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.882526 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.972793 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk"] Oct 02 07:42:59 crc kubenswrapper[4960]: E1002 07:42:59.973325 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fccc04f-086d-4956-b703-9a7615989363" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.973350 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fccc04f-086d-4956-b703-9a7615989363" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.973592 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fccc04f-086d-4956-b703-9a7615989363" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.974374 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.977773 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.977936 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.980241 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.980755 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:42:59 crc kubenswrapper[4960]: I1002 07:42:59.994875 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk"] Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.001062 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.001275 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgm2h\" (UniqueName: \"kubernetes.io/projected/35cd1327-be61-41bc-8cc4-3817a3bea36d-kube-api-access-bgm2h\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.001349 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.102706 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgm2h\" (UniqueName: \"kubernetes.io/projected/35cd1327-be61-41bc-8cc4-3817a3bea36d-kube-api-access-bgm2h\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.102765 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.102823 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.108845 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.108954 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.126411 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgm2h\" (UniqueName: \"kubernetes.io/projected/35cd1327-be61-41bc-8cc4-3817a3bea36d-kube-api-access-bgm2h\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n84sk\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.300876 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.747483 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk"] Oct 02 07:43:00 crc kubenswrapper[4960]: I1002 07:43:00.895713 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" event={"ID":"35cd1327-be61-41bc-8cc4-3817a3bea36d","Type":"ContainerStarted","Data":"31c779f228937a39c842527412dd60041c910ab451bf5643b7fc0624548416a9"} Oct 02 07:43:01 crc kubenswrapper[4960]: I1002 07:43:01.921700 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" event={"ID":"35cd1327-be61-41bc-8cc4-3817a3bea36d","Type":"ContainerStarted","Data":"e043c27ebc4c5a8178b3211d64826b7a69fe8aa119b65f399f0c1deec4a7b3db"} Oct 02 07:43:01 crc kubenswrapper[4960]: I1002 07:43:01.948621 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" podStartSLOduration=2.470327299 podStartE2EDuration="2.948593599s" podCreationTimestamp="2025-10-02 07:42:59 +0000 UTC" firstStartedPulling="2025-10-02 07:43:00.767519439 +0000 UTC m=+1601.799465726" lastFinishedPulling="2025-10-02 07:43:01.245785739 +0000 UTC m=+1602.277732026" observedRunningTime="2025-10-02 07:43:01.945401392 +0000 UTC m=+1602.977347719" watchObservedRunningTime="2025-10-02 07:43:01.948593599 +0000 UTC m=+1602.980539896" Oct 02 07:43:08 crc kubenswrapper[4960]: I1002 07:43:08.330010 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:43:08 crc kubenswrapper[4960]: E1002 07:43:08.332103 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:43:14 crc kubenswrapper[4960]: I1002 07:43:14.074019 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b4f0-account-create-sq74s"] Oct 02 07:43:14 crc kubenswrapper[4960]: I1002 07:43:14.089103 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-de0b-account-create-6q6nn"] Oct 02 07:43:14 crc kubenswrapper[4960]: I1002 07:43:14.098765 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-de0b-account-create-6q6nn"] Oct 02 07:43:14 crc kubenswrapper[4960]: I1002 07:43:14.108791 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b4f0-account-create-sq74s"] Oct 02 07:43:14 crc kubenswrapper[4960]: I1002 07:43:14.341168 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="781f3c80-6287-411d-9c26-ab52d38c8291" path="/var/lib/kubelet/pods/781f3c80-6287-411d-9c26-ab52d38c8291/volumes" Oct 02 07:43:14 crc kubenswrapper[4960]: I1002 07:43:14.341715 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7fa3598-b2f0-4a29-9150-0b881db98bcb" path="/var/lib/kubelet/pods/f7fa3598-b2f0-4a29-9150-0b881db98bcb/volumes" Oct 02 07:43:15 crc kubenswrapper[4960]: I1002 07:43:15.036423 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-sl4rw"] Oct 02 07:43:15 crc kubenswrapper[4960]: I1002 07:43:15.049787 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a856-account-create-h8wpl"] Oct 02 07:43:15 crc kubenswrapper[4960]: I1002 07:43:15.057933 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-sl4rw"] Oct 02 07:43:15 crc kubenswrapper[4960]: I1002 07:43:15.065634 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a856-account-create-h8wpl"] Oct 02 07:43:16 crc kubenswrapper[4960]: I1002 07:43:16.341090 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eb7e0fd-eb02-4f1a-af2f-d61e09104130" path="/var/lib/kubelet/pods/5eb7e0fd-eb02-4f1a-af2f-d61e09104130/volumes" Oct 02 07:43:16 crc kubenswrapper[4960]: I1002 07:43:16.342146 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8478ccab-58bf-45e2-9f7a-c5a4b0bcc216" path="/var/lib/kubelet/pods/8478ccab-58bf-45e2-9f7a-c5a4b0bcc216/volumes" Oct 02 07:43:21 crc kubenswrapper[4960]: I1002 07:43:21.052796 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xjxcq"] Oct 02 07:43:21 crc kubenswrapper[4960]: I1002 07:43:21.065144 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xjxcq"] Oct 02 07:43:22 crc kubenswrapper[4960]: I1002 07:43:22.330052 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:43:22 crc kubenswrapper[4960]: E1002 07:43:22.331355 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:43:22 crc kubenswrapper[4960]: I1002 07:43:22.340229 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14972b11-92c6-4bea-bf0e-f926c18c198f" path="/var/lib/kubelet/pods/14972b11-92c6-4bea-bf0e-f926c18c198f/volumes" Oct 02 07:43:35 crc kubenswrapper[4960]: I1002 07:43:35.331314 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:43:35 crc kubenswrapper[4960]: E1002 07:43:35.332316 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.051806 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-fjshw"] Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.061512 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-fjshw"] Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.263816 4960 scope.go:117] "RemoveContainer" containerID="ebae7086f06dcac78a6692cf60e174260e1fee556e1398b0e38064bd0aab6a51" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.297680 4960 scope.go:117] "RemoveContainer" containerID="1501122a8ea6e8e195ed6a8df5fa7e8886e706c2b22fe0f37b5daacd592784c6" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.344598 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b896e9b5-c55b-4c73-b6d3-36aad785df9f" path="/var/lib/kubelet/pods/b896e9b5-c55b-4c73-b6d3-36aad785df9f/volumes" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.354655 4960 scope.go:117] "RemoveContainer" containerID="76c3a4b0776c2786b3d0129588524cc3d5c3ad291da22aa7fed44e0081b1acf3" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.388905 4960 scope.go:117] "RemoveContainer" containerID="8bb13f6369399e014df1695dbe725e4e2eb136bc25f0b6d74a1833373be25b2a" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.450951 4960 scope.go:117] "RemoveContainer" containerID="7bf1bbac7d2f8d1ba179c0d7907a146854bc10ec869b358abb0682ab64ea5403" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.495648 4960 scope.go:117] "RemoveContainer" containerID="bfde7721b34d159d48b79949ece39ce5e626bb613a808a79f0598c73999c8955" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.546399 4960 scope.go:117] "RemoveContainer" containerID="b1c20644b9ca5b6b0b07d4df04a174a2b4052b9fe8cc4564c0466c18c894e1f5" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.582794 4960 scope.go:117] "RemoveContainer" containerID="95932ea4bc707c32ada6f68dd3182b3dd177a3928dce54959f6136978a031a5f" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.623423 4960 scope.go:117] "RemoveContainer" containerID="d616437001bfc596fdebbbdcf3fbbfff3a2b4e4e733592ed13c3e3511b26d010" Oct 02 07:43:42 crc kubenswrapper[4960]: I1002 07:43:42.676892 4960 scope.go:117] "RemoveContainer" containerID="a495c74eb425cc66ed07cdc3565ac92fb0743250ceb3c81f3a4f528f0697997a" Oct 02 07:43:46 crc kubenswrapper[4960]: I1002 07:43:46.330803 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:43:46 crc kubenswrapper[4960]: E1002 07:43:46.332268 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:43:47 crc kubenswrapper[4960]: I1002 07:43:47.063373 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-x2tp5"] Oct 02 07:43:47 crc kubenswrapper[4960]: I1002 07:43:47.078140 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-x2tp5"] Oct 02 07:43:48 crc kubenswrapper[4960]: I1002 07:43:48.350177 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2b624f-e186-4f60-898b-bce9781f911a" path="/var/lib/kubelet/pods/0f2b624f-e186-4f60-898b-bce9781f911a/volumes" Oct 02 07:43:59 crc kubenswrapper[4960]: I1002 07:43:59.331405 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:43:59 crc kubenswrapper[4960]: E1002 07:43:59.332372 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:44:01 crc kubenswrapper[4960]: I1002 07:44:01.628905 4960 generic.go:334] "Generic (PLEG): container finished" podID="35cd1327-be61-41bc-8cc4-3817a3bea36d" containerID="e043c27ebc4c5a8178b3211d64826b7a69fe8aa119b65f399f0c1deec4a7b3db" exitCode=2 Oct 02 07:44:01 crc kubenswrapper[4960]: I1002 07:44:01.629022 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" event={"ID":"35cd1327-be61-41bc-8cc4-3817a3bea36d","Type":"ContainerDied","Data":"e043c27ebc4c5a8178b3211d64826b7a69fe8aa119b65f399f0c1deec4a7b3db"} Oct 02 07:44:02 crc kubenswrapper[4960]: I1002 07:44:02.043755 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-965xq"] Oct 02 07:44:02 crc kubenswrapper[4960]: I1002 07:44:02.053205 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-965xq"] Oct 02 07:44:02 crc kubenswrapper[4960]: I1002 07:44:02.349519 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f22ba4-0d4e-4012-a82d-b6a98d10b015" path="/var/lib/kubelet/pods/f6f22ba4-0d4e-4012-a82d-b6a98d10b015/volumes" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.084165 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.172779 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgm2h\" (UniqueName: \"kubernetes.io/projected/35cd1327-be61-41bc-8cc4-3817a3bea36d-kube-api-access-bgm2h\") pod \"35cd1327-be61-41bc-8cc4-3817a3bea36d\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.172900 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-inventory\") pod \"35cd1327-be61-41bc-8cc4-3817a3bea36d\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.173002 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-ssh-key\") pod \"35cd1327-be61-41bc-8cc4-3817a3bea36d\" (UID: \"35cd1327-be61-41bc-8cc4-3817a3bea36d\") " Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.180499 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35cd1327-be61-41bc-8cc4-3817a3bea36d-kube-api-access-bgm2h" (OuterVolumeSpecName: "kube-api-access-bgm2h") pod "35cd1327-be61-41bc-8cc4-3817a3bea36d" (UID: "35cd1327-be61-41bc-8cc4-3817a3bea36d"). InnerVolumeSpecName "kube-api-access-bgm2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.201277 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-inventory" (OuterVolumeSpecName: "inventory") pod "35cd1327-be61-41bc-8cc4-3817a3bea36d" (UID: "35cd1327-be61-41bc-8cc4-3817a3bea36d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.219303 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35cd1327-be61-41bc-8cc4-3817a3bea36d" (UID: "35cd1327-be61-41bc-8cc4-3817a3bea36d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.274793 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgm2h\" (UniqueName: \"kubernetes.io/projected/35cd1327-be61-41bc-8cc4-3817a3bea36d-kube-api-access-bgm2h\") on node \"crc\" DevicePath \"\"" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.274833 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.274843 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35cd1327-be61-41bc-8cc4-3817a3bea36d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.655492 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" event={"ID":"35cd1327-be61-41bc-8cc4-3817a3bea36d","Type":"ContainerDied","Data":"31c779f228937a39c842527412dd60041c910ab451bf5643b7fc0624548416a9"} Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.655562 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31c779f228937a39c842527412dd60041c910ab451bf5643b7fc0624548416a9" Oct 02 07:44:03 crc kubenswrapper[4960]: I1002 07:44:03.655609 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk" Oct 02 07:44:10 crc kubenswrapper[4960]: I1002 07:44:10.342575 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:44:10 crc kubenswrapper[4960]: E1002 07:44:10.344589 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.072944 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv"] Oct 02 07:44:11 crc kubenswrapper[4960]: E1002 07:44:11.073743 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35cd1327-be61-41bc-8cc4-3817a3bea36d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.073778 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="35cd1327-be61-41bc-8cc4-3817a3bea36d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.074208 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="35cd1327-be61-41bc-8cc4-3817a3bea36d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.075556 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.079628 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.080161 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.081386 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.081642 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.098302 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv"] Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.172658 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.172809 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.172995 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvjxt\" (UniqueName: \"kubernetes.io/projected/90c6f7d0-fe02-4674-b044-67b7fe197622-kube-api-access-pvjxt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.275905 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.276086 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.276217 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvjxt\" (UniqueName: \"kubernetes.io/projected/90c6f7d0-fe02-4674-b044-67b7fe197622-kube-api-access-pvjxt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.284535 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.285581 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.295661 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvjxt\" (UniqueName: \"kubernetes.io/projected/90c6f7d0-fe02-4674-b044-67b7fe197622-kube-api-access-pvjxt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:11 crc kubenswrapper[4960]: I1002 07:44:11.412324 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:44:12 crc kubenswrapper[4960]: I1002 07:44:12.006153 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv"] Oct 02 07:44:12 crc kubenswrapper[4960]: I1002 07:44:12.767249 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" event={"ID":"90c6f7d0-fe02-4674-b044-67b7fe197622","Type":"ContainerStarted","Data":"6dcbf1fbe7ef5977195f4e9e2aa4245190ae0b77a04f882ee6592ec19d5349b0"} Oct 02 07:44:13 crc kubenswrapper[4960]: I1002 07:44:13.781559 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" event={"ID":"90c6f7d0-fe02-4674-b044-67b7fe197622","Type":"ContainerStarted","Data":"5ec82034dbae0205b17b94d28e4e4ce79f1ddc3ee5f494587eae81b6261251e7"} Oct 02 07:44:13 crc kubenswrapper[4960]: I1002 07:44:13.804576 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" podStartSLOduration=2.141027621 podStartE2EDuration="2.804549397s" podCreationTimestamp="2025-10-02 07:44:11 +0000 UTC" firstStartedPulling="2025-10-02 07:44:12.01650737 +0000 UTC m=+1673.048453657" lastFinishedPulling="2025-10-02 07:44:12.680029146 +0000 UTC m=+1673.711975433" observedRunningTime="2025-10-02 07:44:13.798196687 +0000 UTC m=+1674.830143014" watchObservedRunningTime="2025-10-02 07:44:13.804549397 +0000 UTC m=+1674.836495684" Oct 02 07:44:15 crc kubenswrapper[4960]: I1002 07:44:15.063184 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-gqr95"] Oct 02 07:44:15 crc kubenswrapper[4960]: I1002 07:44:15.079176 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-gqr95"] Oct 02 07:44:15 crc kubenswrapper[4960]: I1002 07:44:15.089724 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-9sml2"] Oct 02 07:44:15 crc kubenswrapper[4960]: I1002 07:44:15.100753 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-9sml2"] Oct 02 07:44:16 crc kubenswrapper[4960]: I1002 07:44:16.346480 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a76d7e-e996-49f6-a3df-7da0d3005b55" path="/var/lib/kubelet/pods/54a76d7e-e996-49f6-a3df-7da0d3005b55/volumes" Oct 02 07:44:16 crc kubenswrapper[4960]: I1002 07:44:16.348339 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a60afc-781b-481f-af68-4d8b9c737d29" path="/var/lib/kubelet/pods/b2a60afc-781b-481f-af68-4d8b9c737d29/volumes" Oct 02 07:44:17 crc kubenswrapper[4960]: I1002 07:44:17.035053 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-tqvz5"] Oct 02 07:44:17 crc kubenswrapper[4960]: I1002 07:44:17.043990 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-tqvz5"] Oct 02 07:44:18 crc kubenswrapper[4960]: I1002 07:44:18.353782 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61aefc93-ba77-48df-af65-af010d8a8efa" path="/var/lib/kubelet/pods/61aefc93-ba77-48df-af65-af010d8a8efa/volumes" Oct 02 07:44:21 crc kubenswrapper[4960]: I1002 07:44:21.329525 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:44:21 crc kubenswrapper[4960]: E1002 07:44:21.330144 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:44:30 crc kubenswrapper[4960]: I1002 07:44:30.048230 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-931c-account-create-t687z"] Oct 02 07:44:30 crc kubenswrapper[4960]: I1002 07:44:30.063612 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-931c-account-create-t687z"] Oct 02 07:44:30 crc kubenswrapper[4960]: I1002 07:44:30.078366 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-2eca-account-create-4knbp"] Oct 02 07:44:30 crc kubenswrapper[4960]: I1002 07:44:30.086493 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-2eca-account-create-4knbp"] Oct 02 07:44:30 crc kubenswrapper[4960]: I1002 07:44:30.355471 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f" path="/var/lib/kubelet/pods/6bb5d0b4-e5a0-4d64-85e1-90eb3c33b61f/volumes" Oct 02 07:44:30 crc kubenswrapper[4960]: I1002 07:44:30.357740 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c40986d9-621e-4cd4-abe0-c70c62baf39b" path="/var/lib/kubelet/pods/c40986d9-621e-4cd4-abe0-c70c62baf39b/volumes" Oct 02 07:44:31 crc kubenswrapper[4960]: I1002 07:44:31.039291 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ca00-account-create-j24m7"] Oct 02 07:44:31 crc kubenswrapper[4960]: I1002 07:44:31.050683 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ca00-account-create-j24m7"] Oct 02 07:44:32 crc kubenswrapper[4960]: I1002 07:44:32.345195 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ad308ec-0722-4c1d-a871-062db558ec6b" path="/var/lib/kubelet/pods/2ad308ec-0722-4c1d-a871-062db558ec6b/volumes" Oct 02 07:44:34 crc kubenswrapper[4960]: I1002 07:44:34.331475 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:44:34 crc kubenswrapper[4960]: E1002 07:44:34.334316 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:44:42 crc kubenswrapper[4960]: I1002 07:44:42.953443 4960 scope.go:117] "RemoveContainer" containerID="6c3207f8696d3f361033f4517227510e73b39529b4b8a2cdd4a0eb78c4149878" Oct 02 07:44:42 crc kubenswrapper[4960]: I1002 07:44:42.988493 4960 scope.go:117] "RemoveContainer" containerID="25f482babd0bd9deb5557faa195fa5669b05f0ec026db0ac52935e8c0c6e0b97" Oct 02 07:44:43 crc kubenswrapper[4960]: I1002 07:44:43.043007 4960 scope.go:117] "RemoveContainer" containerID="51b4bafe895ab6b053d85cf47465eb80c368cba4018eb7c13cd4de98511ea9e4" Oct 02 07:44:43 crc kubenswrapper[4960]: I1002 07:44:43.095270 4960 scope.go:117] "RemoveContainer" containerID="a2f6419062ce4fadb106ed12c86a0b29065f7f8343c471da33d4a96508b0d751" Oct 02 07:44:43 crc kubenswrapper[4960]: I1002 07:44:43.126118 4960 scope.go:117] "RemoveContainer" containerID="33f3ad79cabd4d4be46ab0633313e8253f2c72f73b5090929180955636b941b6" Oct 02 07:44:43 crc kubenswrapper[4960]: I1002 07:44:43.161012 4960 scope.go:117] "RemoveContainer" containerID="3b1a2da3078786d586b932d6f731e9eba3090a4df06071a1bc98faa5f1ceff63" Oct 02 07:44:43 crc kubenswrapper[4960]: I1002 07:44:43.201485 4960 scope.go:117] "RemoveContainer" containerID="177b934ededc13b521097dde3f4552f990e83b874819bd756af4d397b05dec25" Oct 02 07:44:43 crc kubenswrapper[4960]: I1002 07:44:43.239058 4960 scope.go:117] "RemoveContainer" containerID="8eb073a6a9c0d3083222c384c62a22d5fbf2300ab9ee5f7e7bf0347d5a142ed9" Oct 02 07:44:43 crc kubenswrapper[4960]: I1002 07:44:43.259163 4960 scope.go:117] "RemoveContainer" containerID="e1c2490d5b48ed1caf344cc1538ab1d9b5e4534571f04b4ae105ca95646670ee" Oct 02 07:44:47 crc kubenswrapper[4960]: I1002 07:44:47.330112 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:44:47 crc kubenswrapper[4960]: E1002 07:44:47.331145 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:44:52 crc kubenswrapper[4960]: I1002 07:44:52.048712 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-45zxh"] Oct 02 07:44:52 crc kubenswrapper[4960]: I1002 07:44:52.059095 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-45zxh"] Oct 02 07:44:52 crc kubenswrapper[4960]: I1002 07:44:52.343846 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6ef2807-e246-4628-aa45-44ab0805d441" path="/var/lib/kubelet/pods/b6ef2807-e246-4628-aa45-44ab0805d441/volumes" Oct 02 07:44:59 crc kubenswrapper[4960]: I1002 07:44:59.330095 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:44:59 crc kubenswrapper[4960]: E1002 07:44:59.331511 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.169124 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt"] Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.173130 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.186302 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.188327 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.233419 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7wnp\" (UniqueName: \"kubernetes.io/projected/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-kube-api-access-c7wnp\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.233629 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-config-volume\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.234022 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-secret-volume\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.235838 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt"] Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.336099 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7wnp\" (UniqueName: \"kubernetes.io/projected/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-kube-api-access-c7wnp\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.336201 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-config-volume\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.336304 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-secret-volume\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.337435 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-config-volume\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.355094 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-secret-volume\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.365595 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7wnp\" (UniqueName: \"kubernetes.io/projected/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-kube-api-access-c7wnp\") pod \"collect-profiles-29323185-lfrzt\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:00 crc kubenswrapper[4960]: I1002 07:45:00.525517 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:01 crc kubenswrapper[4960]: I1002 07:45:01.074911 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt"] Oct 02 07:45:01 crc kubenswrapper[4960]: I1002 07:45:01.319455 4960 generic.go:334] "Generic (PLEG): container finished" podID="90c6f7d0-fe02-4674-b044-67b7fe197622" containerID="5ec82034dbae0205b17b94d28e4e4ce79f1ddc3ee5f494587eae81b6261251e7" exitCode=0 Oct 02 07:45:01 crc kubenswrapper[4960]: I1002 07:45:01.319537 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" event={"ID":"90c6f7d0-fe02-4674-b044-67b7fe197622","Type":"ContainerDied","Data":"5ec82034dbae0205b17b94d28e4e4ce79f1ddc3ee5f494587eae81b6261251e7"} Oct 02 07:45:01 crc kubenswrapper[4960]: I1002 07:45:01.322491 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" event={"ID":"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f","Type":"ContainerStarted","Data":"fb2feb7439a262cbba4173100dc1ae7ccb25788d5e18bc001697bb1457c4fa70"} Oct 02 07:45:01 crc kubenswrapper[4960]: I1002 07:45:01.322583 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" event={"ID":"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f","Type":"ContainerStarted","Data":"e39482b89944d17f633cb9b42eda91082d57f7ea12bc1078bea2ef386040f044"} Oct 02 07:45:01 crc kubenswrapper[4960]: I1002 07:45:01.370713 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" podStartSLOduration=1.3706835050000001 podStartE2EDuration="1.370683505s" podCreationTimestamp="2025-10-02 07:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:45:01.356348358 +0000 UTC m=+1722.388294645" watchObservedRunningTime="2025-10-02 07:45:01.370683505 +0000 UTC m=+1722.402629802" Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.333282 4960 generic.go:334] "Generic (PLEG): container finished" podID="c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" containerID="fb2feb7439a262cbba4173100dc1ae7ccb25788d5e18bc001697bb1457c4fa70" exitCode=0 Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.343033 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" event={"ID":"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f","Type":"ContainerDied","Data":"fb2feb7439a262cbba4173100dc1ae7ccb25788d5e18bc001697bb1457c4fa70"} Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.824280 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.889129 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvjxt\" (UniqueName: \"kubernetes.io/projected/90c6f7d0-fe02-4674-b044-67b7fe197622-kube-api-access-pvjxt\") pod \"90c6f7d0-fe02-4674-b044-67b7fe197622\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.889289 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-ssh-key\") pod \"90c6f7d0-fe02-4674-b044-67b7fe197622\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.889393 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-inventory\") pod \"90c6f7d0-fe02-4674-b044-67b7fe197622\" (UID: \"90c6f7d0-fe02-4674-b044-67b7fe197622\") " Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.898266 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90c6f7d0-fe02-4674-b044-67b7fe197622-kube-api-access-pvjxt" (OuterVolumeSpecName: "kube-api-access-pvjxt") pod "90c6f7d0-fe02-4674-b044-67b7fe197622" (UID: "90c6f7d0-fe02-4674-b044-67b7fe197622"). InnerVolumeSpecName "kube-api-access-pvjxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.920253 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-inventory" (OuterVolumeSpecName: "inventory") pod "90c6f7d0-fe02-4674-b044-67b7fe197622" (UID: "90c6f7d0-fe02-4674-b044-67b7fe197622"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.931733 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "90c6f7d0-fe02-4674-b044-67b7fe197622" (UID: "90c6f7d0-fe02-4674-b044-67b7fe197622"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.991158 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvjxt\" (UniqueName: \"kubernetes.io/projected/90c6f7d0-fe02-4674-b044-67b7fe197622-kube-api-access-pvjxt\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.991199 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:02 crc kubenswrapper[4960]: I1002 07:45:02.991212 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90c6f7d0-fe02-4674-b044-67b7fe197622-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.350703 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.351030 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv" event={"ID":"90c6f7d0-fe02-4674-b044-67b7fe197622","Type":"ContainerDied","Data":"6dcbf1fbe7ef5977195f4e9e2aa4245190ae0b77a04f882ee6592ec19d5349b0"} Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.351115 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dcbf1fbe7ef5977195f4e9e2aa4245190ae0b77a04f882ee6592ec19d5349b0" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.443232 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6rch2"] Oct 02 07:45:03 crc kubenswrapper[4960]: E1002 07:45:03.443723 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90c6f7d0-fe02-4674-b044-67b7fe197622" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.443748 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="90c6f7d0-fe02-4674-b044-67b7fe197622" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.444020 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="90c6f7d0-fe02-4674-b044-67b7fe197622" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.444768 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.449417 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.450606 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.450778 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.450610 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.490234 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6rch2"] Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.506273 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.506349 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl8v9\" (UniqueName: \"kubernetes.io/projected/d113e843-c488-460c-a156-176eac97fec9-kube-api-access-cl8v9\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.506387 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.607569 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.608175 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl8v9\" (UniqueName: \"kubernetes.io/projected/d113e843-c488-460c-a156-176eac97fec9-kube-api-access-cl8v9\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.608229 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.613087 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.616927 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.626690 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl8v9\" (UniqueName: \"kubernetes.io/projected/d113e843-c488-460c-a156-176eac97fec9-kube-api-access-cl8v9\") pod \"ssh-known-hosts-edpm-deployment-6rch2\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.677196 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.708617 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-secret-volume\") pod \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.708733 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-config-volume\") pod \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.708810 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7wnp\" (UniqueName: \"kubernetes.io/projected/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-kube-api-access-c7wnp\") pod \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\" (UID: \"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f\") " Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.710181 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" (UID: "c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.712764 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" (UID: "c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.713347 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-kube-api-access-c7wnp" (OuterVolumeSpecName: "kube-api-access-c7wnp") pod "c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" (UID: "c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f"). InnerVolumeSpecName "kube-api-access-c7wnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.776835 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.810804 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7wnp\" (UniqueName: \"kubernetes.io/projected/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-kube-api-access-c7wnp\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.810877 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:03 crc kubenswrapper[4960]: I1002 07:45:03.810895 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:04 crc kubenswrapper[4960]: I1002 07:45:04.359558 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6rch2"] Oct 02 07:45:04 crc kubenswrapper[4960]: W1002 07:45:04.361649 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd113e843_c488_460c_a156_176eac97fec9.slice/crio-55e5b53e45e6d2e4262ab6844148295355dd74796e8d7235e1a5aad7ef76b67d WatchSource:0}: Error finding container 55e5b53e45e6d2e4262ab6844148295355dd74796e8d7235e1a5aad7ef76b67d: Status 404 returned error can't find the container with id 55e5b53e45e6d2e4262ab6844148295355dd74796e8d7235e1a5aad7ef76b67d Oct 02 07:45:04 crc kubenswrapper[4960]: I1002 07:45:04.363933 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" event={"ID":"c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f","Type":"ContainerDied","Data":"e39482b89944d17f633cb9b42eda91082d57f7ea12bc1078bea2ef386040f044"} Oct 02 07:45:04 crc kubenswrapper[4960]: I1002 07:45:04.364009 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e39482b89944d17f633cb9b42eda91082d57f7ea12bc1078bea2ef386040f044" Oct 02 07:45:04 crc kubenswrapper[4960]: I1002 07:45:04.364161 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt" Oct 02 07:45:05 crc kubenswrapper[4960]: I1002 07:45:05.383932 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" event={"ID":"d113e843-c488-460c-a156-176eac97fec9","Type":"ContainerStarted","Data":"447c18823fb4d0ce64a5acf74211aaa51b2c00eca2b9c57a74534e00265cbca4"} Oct 02 07:45:05 crc kubenswrapper[4960]: I1002 07:45:05.384487 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" event={"ID":"d113e843-c488-460c-a156-176eac97fec9","Type":"ContainerStarted","Data":"55e5b53e45e6d2e4262ab6844148295355dd74796e8d7235e1a5aad7ef76b67d"} Oct 02 07:45:05 crc kubenswrapper[4960]: I1002 07:45:05.413584 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" podStartSLOduration=1.880122396 podStartE2EDuration="2.41354652s" podCreationTimestamp="2025-10-02 07:45:03 +0000 UTC" firstStartedPulling="2025-10-02 07:45:04.366462453 +0000 UTC m=+1725.398408740" lastFinishedPulling="2025-10-02 07:45:04.899886577 +0000 UTC m=+1725.931832864" observedRunningTime="2025-10-02 07:45:05.410577351 +0000 UTC m=+1726.442523638" watchObservedRunningTime="2025-10-02 07:45:05.41354652 +0000 UTC m=+1726.445492857" Oct 02 07:45:10 crc kubenswrapper[4960]: I1002 07:45:10.342040 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:45:10 crc kubenswrapper[4960]: E1002 07:45:10.343405 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:45:13 crc kubenswrapper[4960]: I1002 07:45:13.485895 4960 generic.go:334] "Generic (PLEG): container finished" podID="d113e843-c488-460c-a156-176eac97fec9" containerID="447c18823fb4d0ce64a5acf74211aaa51b2c00eca2b9c57a74534e00265cbca4" exitCode=0 Oct 02 07:45:13 crc kubenswrapper[4960]: I1002 07:45:13.486026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" event={"ID":"d113e843-c488-460c-a156-176eac97fec9","Type":"ContainerDied","Data":"447c18823fb4d0ce64a5acf74211aaa51b2c00eca2b9c57a74534e00265cbca4"} Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.098576 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.200741 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-inventory-0\") pod \"d113e843-c488-460c-a156-176eac97fec9\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.200885 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-ssh-key-openstack-edpm-ipam\") pod \"d113e843-c488-460c-a156-176eac97fec9\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.200927 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl8v9\" (UniqueName: \"kubernetes.io/projected/d113e843-c488-460c-a156-176eac97fec9-kube-api-access-cl8v9\") pod \"d113e843-c488-460c-a156-176eac97fec9\" (UID: \"d113e843-c488-460c-a156-176eac97fec9\") " Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.208629 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d113e843-c488-460c-a156-176eac97fec9-kube-api-access-cl8v9" (OuterVolumeSpecName: "kube-api-access-cl8v9") pod "d113e843-c488-460c-a156-176eac97fec9" (UID: "d113e843-c488-460c-a156-176eac97fec9"). InnerVolumeSpecName "kube-api-access-cl8v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.230850 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d113e843-c488-460c-a156-176eac97fec9" (UID: "d113e843-c488-460c-a156-176eac97fec9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.246370 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d113e843-c488-460c-a156-176eac97fec9" (UID: "d113e843-c488-460c-a156-176eac97fec9"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.303311 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.303343 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl8v9\" (UniqueName: \"kubernetes.io/projected/d113e843-c488-460c-a156-176eac97fec9-kube-api-access-cl8v9\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.303357 4960 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d113e843-c488-460c-a156-176eac97fec9-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.509938 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" event={"ID":"d113e843-c488-460c-a156-176eac97fec9","Type":"ContainerDied","Data":"55e5b53e45e6d2e4262ab6844148295355dd74796e8d7235e1a5aad7ef76b67d"} Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.510521 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55e5b53e45e6d2e4262ab6844148295355dd74796e8d7235e1a5aad7ef76b67d" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.510495 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6rch2" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.628738 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb"] Oct 02 07:45:15 crc kubenswrapper[4960]: E1002 07:45:15.629239 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" containerName="collect-profiles" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.629258 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" containerName="collect-profiles" Oct 02 07:45:15 crc kubenswrapper[4960]: E1002 07:45:15.629289 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d113e843-c488-460c-a156-176eac97fec9" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.629298 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d113e843-c488-460c-a156-176eac97fec9" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.629492 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" containerName="collect-profiles" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.629529 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d113e843-c488-460c-a156-176eac97fec9" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.630315 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.633316 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.633313 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.633501 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.636232 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.645425 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb"] Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.721750 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkmph\" (UniqueName: \"kubernetes.io/projected/87ba69b2-0ff7-452e-a953-1292fac59ea4-kube-api-access-vkmph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.721841 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.721949 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.824156 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.824417 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkmph\" (UniqueName: \"kubernetes.io/projected/87ba69b2-0ff7-452e-a953-1292fac59ea4-kube-api-access-vkmph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.824472 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.829160 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.830498 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.858080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkmph\" (UniqueName: \"kubernetes.io/projected/87ba69b2-0ff7-452e-a953-1292fac59ea4-kube-api-access-vkmph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-95ngb\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:15 crc kubenswrapper[4960]: I1002 07:45:15.995379 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:16 crc kubenswrapper[4960]: I1002 07:45:16.084652 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-7nd6x"] Oct 02 07:45:16 crc kubenswrapper[4960]: I1002 07:45:16.093894 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-7nd6x"] Oct 02 07:45:16 crc kubenswrapper[4960]: I1002 07:45:16.368557 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8106fcd-0e3e-4440-bbc7-96cd83d369b5" path="/var/lib/kubelet/pods/b8106fcd-0e3e-4440-bbc7-96cd83d369b5/volumes" Oct 02 07:45:16 crc kubenswrapper[4960]: I1002 07:45:16.653643 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb"] Oct 02 07:45:17 crc kubenswrapper[4960]: I1002 07:45:17.540419 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" event={"ID":"87ba69b2-0ff7-452e-a953-1292fac59ea4","Type":"ContainerStarted","Data":"2bd9cb016a0bd6706ec701b199c65be55869d511eed34e54b9a3f245df73d579"} Oct 02 07:45:17 crc kubenswrapper[4960]: I1002 07:45:17.540937 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" event={"ID":"87ba69b2-0ff7-452e-a953-1292fac59ea4","Type":"ContainerStarted","Data":"3d4d40d78e239b00faa551301e2b42b05e0913832da6f8504f36c4654a4c9761"} Oct 02 07:45:17 crc kubenswrapper[4960]: I1002 07:45:17.566176 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" podStartSLOduration=2.0834392250000002 podStartE2EDuration="2.566156056s" podCreationTimestamp="2025-10-02 07:45:15 +0000 UTC" firstStartedPulling="2025-10-02 07:45:16.651499491 +0000 UTC m=+1737.683445798" lastFinishedPulling="2025-10-02 07:45:17.134216322 +0000 UTC m=+1738.166162629" observedRunningTime="2025-10-02 07:45:17.563168076 +0000 UTC m=+1738.595114363" watchObservedRunningTime="2025-10-02 07:45:17.566156056 +0000 UTC m=+1738.598102343" Oct 02 07:45:21 crc kubenswrapper[4960]: I1002 07:45:21.044142 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zxzpz"] Oct 02 07:45:21 crc kubenswrapper[4960]: I1002 07:45:21.064382 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zxzpz"] Oct 02 07:45:22 crc kubenswrapper[4960]: I1002 07:45:22.341494 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="334bd1b9-aae9-4744-b800-817bbf01d3d9" path="/var/lib/kubelet/pods/334bd1b9-aae9-4744-b800-817bbf01d3d9/volumes" Oct 02 07:45:25 crc kubenswrapper[4960]: I1002 07:45:25.330082 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:45:25 crc kubenswrapper[4960]: E1002 07:45:25.332139 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:45:25 crc kubenswrapper[4960]: I1002 07:45:25.632171 4960 generic.go:334] "Generic (PLEG): container finished" podID="87ba69b2-0ff7-452e-a953-1292fac59ea4" containerID="2bd9cb016a0bd6706ec701b199c65be55869d511eed34e54b9a3f245df73d579" exitCode=0 Oct 02 07:45:25 crc kubenswrapper[4960]: I1002 07:45:25.632264 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" event={"ID":"87ba69b2-0ff7-452e-a953-1292fac59ea4","Type":"ContainerDied","Data":"2bd9cb016a0bd6706ec701b199c65be55869d511eed34e54b9a3f245df73d579"} Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.199936 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.315617 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkmph\" (UniqueName: \"kubernetes.io/projected/87ba69b2-0ff7-452e-a953-1292fac59ea4-kube-api-access-vkmph\") pod \"87ba69b2-0ff7-452e-a953-1292fac59ea4\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.315676 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-inventory\") pod \"87ba69b2-0ff7-452e-a953-1292fac59ea4\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.315719 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-ssh-key\") pod \"87ba69b2-0ff7-452e-a953-1292fac59ea4\" (UID: \"87ba69b2-0ff7-452e-a953-1292fac59ea4\") " Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.325076 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ba69b2-0ff7-452e-a953-1292fac59ea4-kube-api-access-vkmph" (OuterVolumeSpecName: "kube-api-access-vkmph") pod "87ba69b2-0ff7-452e-a953-1292fac59ea4" (UID: "87ba69b2-0ff7-452e-a953-1292fac59ea4"). InnerVolumeSpecName "kube-api-access-vkmph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.352180 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-inventory" (OuterVolumeSpecName: "inventory") pod "87ba69b2-0ff7-452e-a953-1292fac59ea4" (UID: "87ba69b2-0ff7-452e-a953-1292fac59ea4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.352651 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87ba69b2-0ff7-452e-a953-1292fac59ea4" (UID: "87ba69b2-0ff7-452e-a953-1292fac59ea4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.419341 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkmph\" (UniqueName: \"kubernetes.io/projected/87ba69b2-0ff7-452e-a953-1292fac59ea4-kube-api-access-vkmph\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.419671 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.419755 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87ba69b2-0ff7-452e-a953-1292fac59ea4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.654741 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" event={"ID":"87ba69b2-0ff7-452e-a953-1292fac59ea4","Type":"ContainerDied","Data":"3d4d40d78e239b00faa551301e2b42b05e0913832da6f8504f36c4654a4c9761"} Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.654794 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d4d40d78e239b00faa551301e2b42b05e0913832da6f8504f36c4654a4c9761" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.654889 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.756994 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2"] Oct 02 07:45:27 crc kubenswrapper[4960]: E1002 07:45:27.757506 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ba69b2-0ff7-452e-a953-1292fac59ea4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.757535 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ba69b2-0ff7-452e-a953-1292fac59ea4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.757772 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ba69b2-0ff7-452e-a953-1292fac59ea4" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.758724 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.761745 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.761809 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.761745 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.761883 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.774278 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2"] Oct 02 07:45:27 crc kubenswrapper[4960]: E1002 07:45:27.908145 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87ba69b2_0ff7_452e_a953_1292fac59ea4.slice/crio-3d4d40d78e239b00faa551301e2b42b05e0913832da6f8504f36c4654a4c9761\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87ba69b2_0ff7_452e_a953_1292fac59ea4.slice\": RecentStats: unable to find data in memory cache]" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.929889 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.930044 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdwzl\" (UniqueName: \"kubernetes.io/projected/2bbd3f4c-233a-41b8-961b-79a07143dae3-kube-api-access-kdwzl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:27 crc kubenswrapper[4960]: I1002 07:45:27.930104 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.032296 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdwzl\" (UniqueName: \"kubernetes.io/projected/2bbd3f4c-233a-41b8-961b-79a07143dae3-kube-api-access-kdwzl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.032410 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.032510 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.040010 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.047247 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.055390 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdwzl\" (UniqueName: \"kubernetes.io/projected/2bbd3f4c-233a-41b8-961b-79a07143dae3-kube-api-access-kdwzl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.091445 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.646317 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2"] Oct 02 07:45:28 crc kubenswrapper[4960]: I1002 07:45:28.667072 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" event={"ID":"2bbd3f4c-233a-41b8-961b-79a07143dae3","Type":"ContainerStarted","Data":"50b58439b11d842087bf90f8cbdfe8c3c38242527baaad8f398776d3778a398e"} Oct 02 07:45:29 crc kubenswrapper[4960]: I1002 07:45:29.680350 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" event={"ID":"2bbd3f4c-233a-41b8-961b-79a07143dae3","Type":"ContainerStarted","Data":"0795d5125ef89562e2e96210224e40f0cc4271e33a229b9ad06d52c9c206acb5"} Oct 02 07:45:38 crc kubenswrapper[4960]: I1002 07:45:38.330676 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:45:38 crc kubenswrapper[4960]: E1002 07:45:38.332455 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:45:38 crc kubenswrapper[4960]: I1002 07:45:38.811160 4960 generic.go:334] "Generic (PLEG): container finished" podID="2bbd3f4c-233a-41b8-961b-79a07143dae3" containerID="0795d5125ef89562e2e96210224e40f0cc4271e33a229b9ad06d52c9c206acb5" exitCode=0 Oct 02 07:45:38 crc kubenswrapper[4960]: I1002 07:45:38.811241 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" event={"ID":"2bbd3f4c-233a-41b8-961b-79a07143dae3","Type":"ContainerDied","Data":"0795d5125ef89562e2e96210224e40f0cc4271e33a229b9ad06d52c9c206acb5"} Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.255420 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.442806 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-inventory\") pod \"2bbd3f4c-233a-41b8-961b-79a07143dae3\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.442899 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdwzl\" (UniqueName: \"kubernetes.io/projected/2bbd3f4c-233a-41b8-961b-79a07143dae3-kube-api-access-kdwzl\") pod \"2bbd3f4c-233a-41b8-961b-79a07143dae3\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.443266 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-ssh-key\") pod \"2bbd3f4c-233a-41b8-961b-79a07143dae3\" (UID: \"2bbd3f4c-233a-41b8-961b-79a07143dae3\") " Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.452313 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbd3f4c-233a-41b8-961b-79a07143dae3-kube-api-access-kdwzl" (OuterVolumeSpecName: "kube-api-access-kdwzl") pod "2bbd3f4c-233a-41b8-961b-79a07143dae3" (UID: "2bbd3f4c-233a-41b8-961b-79a07143dae3"). InnerVolumeSpecName "kube-api-access-kdwzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.474626 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-inventory" (OuterVolumeSpecName: "inventory") pod "2bbd3f4c-233a-41b8-961b-79a07143dae3" (UID: "2bbd3f4c-233a-41b8-961b-79a07143dae3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.476338 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2bbd3f4c-233a-41b8-961b-79a07143dae3" (UID: "2bbd3f4c-233a-41b8-961b-79a07143dae3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.546219 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.546271 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdwzl\" (UniqueName: \"kubernetes.io/projected/2bbd3f4c-233a-41b8-961b-79a07143dae3-kube-api-access-kdwzl\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.546289 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbd3f4c-233a-41b8-961b-79a07143dae3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.837178 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" event={"ID":"2bbd3f4c-233a-41b8-961b-79a07143dae3","Type":"ContainerDied","Data":"50b58439b11d842087bf90f8cbdfe8c3c38242527baaad8f398776d3778a398e"} Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.837612 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50b58439b11d842087bf90f8cbdfe8c3c38242527baaad8f398776d3778a398e" Oct 02 07:45:40 crc kubenswrapper[4960]: I1002 07:45:40.837356 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2" Oct 02 07:45:43 crc kubenswrapper[4960]: I1002 07:45:43.433761 4960 scope.go:117] "RemoveContainer" containerID="cfcb3e868db40531716a774898dfaab314ca1fc0e1bf24fef47df76f39fcf6d0" Oct 02 07:45:43 crc kubenswrapper[4960]: I1002 07:45:43.473375 4960 scope.go:117] "RemoveContainer" containerID="6389c63ca445a838d188e1446238f1d73f34593d1ba53928214879d4167919c4" Oct 02 07:45:43 crc kubenswrapper[4960]: I1002 07:45:43.548761 4960 scope.go:117] "RemoveContainer" containerID="6da09db331c50808dd892510883439a1650cbd3d3e4951a9521f62a22380f5a0" Oct 02 07:45:49 crc kubenswrapper[4960]: I1002 07:45:49.330593 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:45:49 crc kubenswrapper[4960]: E1002 07:45:49.331498 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:46:02 crc kubenswrapper[4960]: I1002 07:46:02.086307 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-k5544"] Oct 02 07:46:02 crc kubenswrapper[4960]: I1002 07:46:02.094440 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-k5544"] Oct 02 07:46:02 crc kubenswrapper[4960]: I1002 07:46:02.343413 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dff90544-5bf1-4efd-becf-7f5f271f5de6" path="/var/lib/kubelet/pods/dff90544-5bf1-4efd-becf-7f5f271f5de6/volumes" Oct 02 07:46:04 crc kubenswrapper[4960]: I1002 07:46:04.330866 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:46:04 crc kubenswrapper[4960]: E1002 07:46:04.331872 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:46:16 crc kubenswrapper[4960]: I1002 07:46:16.330919 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:46:16 crc kubenswrapper[4960]: E1002 07:46:16.332277 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:46:31 crc kubenswrapper[4960]: I1002 07:46:31.330364 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:46:31 crc kubenswrapper[4960]: E1002 07:46:31.331256 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:46:43 crc kubenswrapper[4960]: I1002 07:46:43.751923 4960 scope.go:117] "RemoveContainer" containerID="315df6ccfeccbdb10e579ff9ab235bfeb9cb06d2a0b0bc10bae78531979b53e8" Oct 02 07:46:45 crc kubenswrapper[4960]: I1002 07:46:45.333626 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:46:45 crc kubenswrapper[4960]: E1002 07:46:45.334366 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:46:56 crc kubenswrapper[4960]: I1002 07:46:56.333273 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:46:56 crc kubenswrapper[4960]: E1002 07:46:56.334812 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:47:07 crc kubenswrapper[4960]: I1002 07:47:07.332127 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:47:07 crc kubenswrapper[4960]: I1002 07:47:07.827560 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"8719c408af0adf4927f760c6fbcd579ca45bc4eb06efea39eb9e845e3a7c5024"} Oct 02 07:49:29 crc kubenswrapper[4960]: I1002 07:49:29.149768 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:49:29 crc kubenswrapper[4960]: I1002 07:49:29.150764 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:49:32 crc kubenswrapper[4960]: E1002 07:49:32.802827 4960 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.177:38870->38.102.83.177:37215: write tcp 38.102.83.177:38870->38.102.83.177:37215: write: broken pipe Oct 02 07:49:59 crc kubenswrapper[4960]: I1002 07:49:59.150664 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:49:59 crc kubenswrapper[4960]: I1002 07:49:59.151761 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.435146 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.454095 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-7r8kk"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.470324 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.477043 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.482667 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.492485 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.501047 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.509120 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.518741 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.525964 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6rch2"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.533787 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.539913 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.546514 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-48z92"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.552363 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrwz6"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.558466 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tt5pv"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.567456 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-95ngb"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.578720 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rq2k2"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.597400 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n84sk"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.619856 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6rch2"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.631068 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4vv2b"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.642344 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8s4kl"] Oct 02 07:50:18 crc kubenswrapper[4960]: I1002 07:50:18.651963 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f2m4x"] Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.344472 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bbd3f4c-233a-41b8-961b-79a07143dae3" path="/var/lib/kubelet/pods/2bbd3f4c-233a-41b8-961b-79a07143dae3/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.346223 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35cd1327-be61-41bc-8cc4-3817a3bea36d" path="/var/lib/kubelet/pods/35cd1327-be61-41bc-8cc4-3817a3bea36d/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.347446 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4077b2fa-ed35-41af-bca4-562ff6d9c100" path="/var/lib/kubelet/pods/4077b2fa-ed35-41af-bca4-562ff6d9c100/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.348634 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44e92992-9dec-4817-a070-035bde48c556" path="/var/lib/kubelet/pods/44e92992-9dec-4817-a070-035bde48c556/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.351966 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fccc04f-086d-4956-b703-9a7615989363" path="/var/lib/kubelet/pods/4fccc04f-086d-4956-b703-9a7615989363/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.354068 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dac1c0f-0383-4a34-a3b9-10dbd82dbd05" path="/var/lib/kubelet/pods/7dac1c0f-0383-4a34-a3b9-10dbd82dbd05/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.355549 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ba69b2-0ff7-452e-a953-1292fac59ea4" path="/var/lib/kubelet/pods/87ba69b2-0ff7-452e-a953-1292fac59ea4/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.356547 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="881f2dfd-9294-444d-9c05-18b72a9128a2" path="/var/lib/kubelet/pods/881f2dfd-9294-444d-9c05-18b72a9128a2/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.357536 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90c6f7d0-fe02-4674-b044-67b7fe197622" path="/var/lib/kubelet/pods/90c6f7d0-fe02-4674-b044-67b7fe197622/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.358897 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d113e843-c488-460c-a156-176eac97fec9" path="/var/lib/kubelet/pods/d113e843-c488-460c-a156-176eac97fec9/volumes" Oct 02 07:50:20 crc kubenswrapper[4960]: I1002 07:50:20.359446 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9302552-a6d8-43f6-93d7-05db5a3a9e30" path="/var/lib/kubelet/pods/f9302552-a6d8-43f6-93d7-05db5a3a9e30/volumes" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.288263 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-84l85"] Oct 02 07:50:22 crc kubenswrapper[4960]: E1002 07:50:22.289335 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbd3f4c-233a-41b8-961b-79a07143dae3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.289363 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbd3f4c-233a-41b8-961b-79a07143dae3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.289750 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbd3f4c-233a-41b8-961b-79a07143dae3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.294490 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.307306 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-84l85"] Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.476309 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-utilities\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.476955 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-catalog-content\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.477153 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jcrg\" (UniqueName: \"kubernetes.io/projected/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-kube-api-access-7jcrg\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.579536 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-catalog-content\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.580086 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jcrg\" (UniqueName: \"kubernetes.io/projected/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-kube-api-access-7jcrg\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.580194 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-utilities\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.580706 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-catalog-content\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.580954 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-utilities\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.604860 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jcrg\" (UniqueName: \"kubernetes.io/projected/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-kube-api-access-7jcrg\") pod \"redhat-marketplace-84l85\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:22 crc kubenswrapper[4960]: I1002 07:50:22.637571 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:23 crc kubenswrapper[4960]: I1002 07:50:23.204264 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-84l85"] Oct 02 07:50:23 crc kubenswrapper[4960]: I1002 07:50:23.798633 4960 generic.go:334] "Generic (PLEG): container finished" podID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerID="83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84" exitCode=0 Oct 02 07:50:23 crc kubenswrapper[4960]: I1002 07:50:23.798768 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84l85" event={"ID":"712f14be-9dea-4b07-9bd6-72c90fbcb7c7","Type":"ContainerDied","Data":"83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84"} Oct 02 07:50:23 crc kubenswrapper[4960]: I1002 07:50:23.799312 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84l85" event={"ID":"712f14be-9dea-4b07-9bd6-72c90fbcb7c7","Type":"ContainerStarted","Data":"adbcf1a107f18458392592b8ddbd89fae2849449a589bfe01082791ab4d8b147"} Oct 02 07:50:23 crc kubenswrapper[4960]: I1002 07:50:23.802299 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.745987 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w"] Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.749802 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.754414 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.754692 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.754691 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.757799 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.759198 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.769490 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w"] Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.869107 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk4q5\" (UniqueName: \"kubernetes.io/projected/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-kube-api-access-nk4q5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.869191 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.869516 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.869785 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.869822 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.971999 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.972064 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.972131 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk4q5\" (UniqueName: \"kubernetes.io/projected/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-kube-api-access-nk4q5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.972242 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.972449 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.981863 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.981845 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.982260 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.987717 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:24 crc kubenswrapper[4960]: I1002 07:50:24.993204 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk4q5\" (UniqueName: \"kubernetes.io/projected/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-kube-api-access-nk4q5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:25 crc kubenswrapper[4960]: I1002 07:50:25.086373 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:25 crc kubenswrapper[4960]: I1002 07:50:25.638661 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w"] Oct 02 07:50:25 crc kubenswrapper[4960]: W1002 07:50:25.651548 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode621ea76_77e5_4ae4_aee0_eb92a4f247a5.slice/crio-907f5eccc18f4e0460937395f93bbd68f3cbbc0ac799ee26c1e63a4df684b462 WatchSource:0}: Error finding container 907f5eccc18f4e0460937395f93bbd68f3cbbc0ac799ee26c1e63a4df684b462: Status 404 returned error can't find the container with id 907f5eccc18f4e0460937395f93bbd68f3cbbc0ac799ee26c1e63a4df684b462 Oct 02 07:50:25 crc kubenswrapper[4960]: I1002 07:50:25.822218 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" event={"ID":"e621ea76-77e5-4ae4-aee0-eb92a4f247a5","Type":"ContainerStarted","Data":"907f5eccc18f4e0460937395f93bbd68f3cbbc0ac799ee26c1e63a4df684b462"} Oct 02 07:50:25 crc kubenswrapper[4960]: I1002 07:50:25.828386 4960 generic.go:334] "Generic (PLEG): container finished" podID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerID="2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262" exitCode=0 Oct 02 07:50:25 crc kubenswrapper[4960]: I1002 07:50:25.828489 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84l85" event={"ID":"712f14be-9dea-4b07-9bd6-72c90fbcb7c7","Type":"ContainerDied","Data":"2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262"} Oct 02 07:50:26 crc kubenswrapper[4960]: I1002 07:50:26.850865 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" event={"ID":"e621ea76-77e5-4ae4-aee0-eb92a4f247a5","Type":"ContainerStarted","Data":"2ed556744a48424cb4cab23102d51e53ebc9aced344c756a9a6273d89fee2864"} Oct 02 07:50:26 crc kubenswrapper[4960]: I1002 07:50:26.860295 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84l85" event={"ID":"712f14be-9dea-4b07-9bd6-72c90fbcb7c7","Type":"ContainerStarted","Data":"020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b"} Oct 02 07:50:26 crc kubenswrapper[4960]: I1002 07:50:26.896639 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-84l85" podStartSLOduration=2.421687701 podStartE2EDuration="4.896617051s" podCreationTimestamp="2025-10-02 07:50:22 +0000 UTC" firstStartedPulling="2025-10-02 07:50:23.801928335 +0000 UTC m=+2044.833874632" lastFinishedPulling="2025-10-02 07:50:26.276857695 +0000 UTC m=+2047.308803982" observedRunningTime="2025-10-02 07:50:26.894678501 +0000 UTC m=+2047.926624798" watchObservedRunningTime="2025-10-02 07:50:26.896617051 +0000 UTC m=+2047.928563338" Oct 02 07:50:26 crc kubenswrapper[4960]: I1002 07:50:26.906114 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" podStartSLOduration=2.24411207 podStartE2EDuration="2.906089536s" podCreationTimestamp="2025-10-02 07:50:24 +0000 UTC" firstStartedPulling="2025-10-02 07:50:25.654765478 +0000 UTC m=+2046.686711765" lastFinishedPulling="2025-10-02 07:50:26.316742944 +0000 UTC m=+2047.348689231" observedRunningTime="2025-10-02 07:50:26.870931288 +0000 UTC m=+2047.902877595" watchObservedRunningTime="2025-10-02 07:50:26.906089536 +0000 UTC m=+2047.938035823" Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.150396 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.150471 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.150526 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.151359 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8719c408af0adf4927f760c6fbcd579ca45bc4eb06efea39eb9e845e3a7c5024"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.151421 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://8719c408af0adf4927f760c6fbcd579ca45bc4eb06efea39eb9e845e3a7c5024" gracePeriod=600 Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.891828 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="8719c408af0adf4927f760c6fbcd579ca45bc4eb06efea39eb9e845e3a7c5024" exitCode=0 Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.891903 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"8719c408af0adf4927f760c6fbcd579ca45bc4eb06efea39eb9e845e3a7c5024"} Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.892635 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755"} Oct 02 07:50:29 crc kubenswrapper[4960]: I1002 07:50:29.892678 4960 scope.go:117] "RemoveContainer" containerID="43a8256ddb3ae3a3e7b8f026f399048078fe7eec7d85c3b2987733a6c89a5678" Oct 02 07:50:32 crc kubenswrapper[4960]: I1002 07:50:32.638649 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:32 crc kubenswrapper[4960]: I1002 07:50:32.639043 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:32 crc kubenswrapper[4960]: I1002 07:50:32.705009 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:32 crc kubenswrapper[4960]: I1002 07:50:32.983924 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:33 crc kubenswrapper[4960]: I1002 07:50:33.048393 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-84l85"] Oct 02 07:50:34 crc kubenswrapper[4960]: I1002 07:50:34.960356 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-84l85" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="registry-server" containerID="cri-o://020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b" gracePeriod=2 Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.450443 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.514323 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-utilities\") pod \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.514695 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jcrg\" (UniqueName: \"kubernetes.io/projected/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-kube-api-access-7jcrg\") pod \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.514945 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-catalog-content\") pod \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\" (UID: \"712f14be-9dea-4b07-9bd6-72c90fbcb7c7\") " Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.515468 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-utilities" (OuterVolumeSpecName: "utilities") pod "712f14be-9dea-4b07-9bd6-72c90fbcb7c7" (UID: "712f14be-9dea-4b07-9bd6-72c90fbcb7c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.515924 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.521781 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-kube-api-access-7jcrg" (OuterVolumeSpecName: "kube-api-access-7jcrg") pod "712f14be-9dea-4b07-9bd6-72c90fbcb7c7" (UID: "712f14be-9dea-4b07-9bd6-72c90fbcb7c7"). InnerVolumeSpecName "kube-api-access-7jcrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.532760 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "712f14be-9dea-4b07-9bd6-72c90fbcb7c7" (UID: "712f14be-9dea-4b07-9bd6-72c90fbcb7c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.617746 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jcrg\" (UniqueName: \"kubernetes.io/projected/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-kube-api-access-7jcrg\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.617824 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712f14be-9dea-4b07-9bd6-72c90fbcb7c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.978537 4960 generic.go:334] "Generic (PLEG): container finished" podID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerID="020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b" exitCode=0 Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.978625 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84l85" event={"ID":"712f14be-9dea-4b07-9bd6-72c90fbcb7c7","Type":"ContainerDied","Data":"020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b"} Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.979129 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84l85" event={"ID":"712f14be-9dea-4b07-9bd6-72c90fbcb7c7","Type":"ContainerDied","Data":"adbcf1a107f18458392592b8ddbd89fae2849449a589bfe01082791ab4d8b147"} Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.979166 4960 scope.go:117] "RemoveContainer" containerID="020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b" Oct 02 07:50:35 crc kubenswrapper[4960]: I1002 07:50:35.978754 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84l85" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.011113 4960 scope.go:117] "RemoveContainer" containerID="2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.035881 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-84l85"] Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.045207 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-84l85"] Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.053359 4960 scope.go:117] "RemoveContainer" containerID="83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.095034 4960 scope.go:117] "RemoveContainer" containerID="020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b" Oct 02 07:50:36 crc kubenswrapper[4960]: E1002 07:50:36.095664 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b\": container with ID starting with 020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b not found: ID does not exist" containerID="020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.095715 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b"} err="failed to get container status \"020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b\": rpc error: code = NotFound desc = could not find container \"020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b\": container with ID starting with 020811390e3b87e3d3b77b3c4f736b68cf69aa13afe46a5000742bd077b8e97b not found: ID does not exist" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.095751 4960 scope.go:117] "RemoveContainer" containerID="2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262" Oct 02 07:50:36 crc kubenswrapper[4960]: E1002 07:50:36.096330 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262\": container with ID starting with 2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262 not found: ID does not exist" containerID="2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.096367 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262"} err="failed to get container status \"2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262\": rpc error: code = NotFound desc = could not find container \"2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262\": container with ID starting with 2d729498c2d8dc4211231330a30a82189622771b4ce06d38b95cfdfc1e229262 not found: ID does not exist" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.096388 4960 scope.go:117] "RemoveContainer" containerID="83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84" Oct 02 07:50:36 crc kubenswrapper[4960]: E1002 07:50:36.096721 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84\": container with ID starting with 83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84 not found: ID does not exist" containerID="83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.096762 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84"} err="failed to get container status \"83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84\": rpc error: code = NotFound desc = could not find container \"83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84\": container with ID starting with 83d8c7fe3228c974315a3b705e68727e67b644e43cdcadc3047e3b7d06654c84 not found: ID does not exist" Oct 02 07:50:36 crc kubenswrapper[4960]: I1002 07:50:36.349217 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" path="/var/lib/kubelet/pods/712f14be-9dea-4b07-9bd6-72c90fbcb7c7/volumes" Oct 02 07:50:38 crc kubenswrapper[4960]: I1002 07:50:38.000904 4960 generic.go:334] "Generic (PLEG): container finished" podID="e621ea76-77e5-4ae4-aee0-eb92a4f247a5" containerID="2ed556744a48424cb4cab23102d51e53ebc9aced344c756a9a6273d89fee2864" exitCode=0 Oct 02 07:50:38 crc kubenswrapper[4960]: I1002 07:50:38.001036 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" event={"ID":"e621ea76-77e5-4ae4-aee0-eb92a4f247a5","Type":"ContainerDied","Data":"2ed556744a48424cb4cab23102d51e53ebc9aced344c756a9a6273d89fee2864"} Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.547600 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.603538 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-repo-setup-combined-ca-bundle\") pod \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.603670 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-inventory\") pod \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.603753 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ssh-key\") pod \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.603778 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ceph\") pod \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.604002 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk4q5\" (UniqueName: \"kubernetes.io/projected/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-kube-api-access-nk4q5\") pod \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\" (UID: \"e621ea76-77e5-4ae4-aee0-eb92a4f247a5\") " Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.612561 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-kube-api-access-nk4q5" (OuterVolumeSpecName: "kube-api-access-nk4q5") pod "e621ea76-77e5-4ae4-aee0-eb92a4f247a5" (UID: "e621ea76-77e5-4ae4-aee0-eb92a4f247a5"). InnerVolumeSpecName "kube-api-access-nk4q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.614395 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ceph" (OuterVolumeSpecName: "ceph") pod "e621ea76-77e5-4ae4-aee0-eb92a4f247a5" (UID: "e621ea76-77e5-4ae4-aee0-eb92a4f247a5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.635162 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e621ea76-77e5-4ae4-aee0-eb92a4f247a5" (UID: "e621ea76-77e5-4ae4-aee0-eb92a4f247a5"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.652693 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-inventory" (OuterVolumeSpecName: "inventory") pod "e621ea76-77e5-4ae4-aee0-eb92a4f247a5" (UID: "e621ea76-77e5-4ae4-aee0-eb92a4f247a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.653112 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e621ea76-77e5-4ae4-aee0-eb92a4f247a5" (UID: "e621ea76-77e5-4ae4-aee0-eb92a4f247a5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.707489 4960 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.707868 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.707885 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.707895 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:39 crc kubenswrapper[4960]: I1002 07:50:39.707906 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk4q5\" (UniqueName: \"kubernetes.io/projected/e621ea76-77e5-4ae4-aee0-eb92a4f247a5-kube-api-access-nk4q5\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.050662 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" event={"ID":"e621ea76-77e5-4ae4-aee0-eb92a4f247a5","Type":"ContainerDied","Data":"907f5eccc18f4e0460937395f93bbd68f3cbbc0ac799ee26c1e63a4df684b462"} Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.051065 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="907f5eccc18f4e0460937395f93bbd68f3cbbc0ac799ee26c1e63a4df684b462" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.050767 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.137657 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w"] Oct 02 07:50:40 crc kubenswrapper[4960]: E1002 07:50:40.138631 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="extract-utilities" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.138655 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="extract-utilities" Oct 02 07:50:40 crc kubenswrapper[4960]: E1002 07:50:40.138689 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e621ea76-77e5-4ae4-aee0-eb92a4f247a5" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.138700 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e621ea76-77e5-4ae4-aee0-eb92a4f247a5" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:40 crc kubenswrapper[4960]: E1002 07:50:40.138722 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="registry-server" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.138731 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="registry-server" Oct 02 07:50:40 crc kubenswrapper[4960]: E1002 07:50:40.138744 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="extract-content" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.138751 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="extract-content" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.138961 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e621ea76-77e5-4ae4-aee0-eb92a4f247a5" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.139008 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="712f14be-9dea-4b07-9bd6-72c90fbcb7c7" containerName="registry-server" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.139791 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.142823 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.143042 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.143661 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.144576 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.150621 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.151864 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w"] Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.219204 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.219264 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.219299 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2s6z\" (UniqueName: \"kubernetes.io/projected/9d905943-1348-47e8-81a5-5794bb84f7ff-kube-api-access-b2s6z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.219362 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.220225 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.323591 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.323724 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.323767 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.323812 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2s6z\" (UniqueName: \"kubernetes.io/projected/9d905943-1348-47e8-81a5-5794bb84f7ff-kube-api-access-b2s6z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.323910 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.330091 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.331097 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.332434 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.345811 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.357748 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2s6z\" (UniqueName: \"kubernetes.io/projected/9d905943-1348-47e8-81a5-5794bb84f7ff-kube-api-access-b2s6z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:40 crc kubenswrapper[4960]: I1002 07:50:40.458555 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:50:41 crc kubenswrapper[4960]: I1002 07:50:41.048167 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w"] Oct 02 07:50:42 crc kubenswrapper[4960]: I1002 07:50:42.076011 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" event={"ID":"9d905943-1348-47e8-81a5-5794bb84f7ff","Type":"ContainerStarted","Data":"e776d0605b6f612bd9f1a46dabae5611a6d53b5befdf1b53d96698f94961b456"} Oct 02 07:50:43 crc kubenswrapper[4960]: I1002 07:50:43.095495 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" event={"ID":"9d905943-1348-47e8-81a5-5794bb84f7ff","Type":"ContainerStarted","Data":"4ad048516b2480ea54f8806270a893d7907b7280e086eb597f2e389b416fa3c3"} Oct 02 07:50:43 crc kubenswrapper[4960]: I1002 07:50:43.129806 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" podStartSLOduration=2.403548501 podStartE2EDuration="3.129771404s" podCreationTimestamp="2025-10-02 07:50:40 +0000 UTC" firstStartedPulling="2025-10-02 07:50:41.075109882 +0000 UTC m=+2062.107056169" lastFinishedPulling="2025-10-02 07:50:41.801332775 +0000 UTC m=+2062.833279072" observedRunningTime="2025-10-02 07:50:43.118651847 +0000 UTC m=+2064.150598124" watchObservedRunningTime="2025-10-02 07:50:43.129771404 +0000 UTC m=+2064.161717751" Oct 02 07:50:43 crc kubenswrapper[4960]: I1002 07:50:43.908810 4960 scope.go:117] "RemoveContainer" containerID="9da040311f074cb1188e386be5749e2a92e4195f811a6efcac29e6f5ed23f9f9" Oct 02 07:50:43 crc kubenswrapper[4960]: I1002 07:50:43.947796 4960 scope.go:117] "RemoveContainer" containerID="e043c27ebc4c5a8178b3211d64826b7a69fe8aa119b65f399f0c1deec4a7b3db" Oct 02 07:50:44 crc kubenswrapper[4960]: I1002 07:50:44.002489 4960 scope.go:117] "RemoveContainer" containerID="5ec82034dbae0205b17b94d28e4e4ce79f1ddc3ee5f494587eae81b6261251e7" Oct 02 07:50:44 crc kubenswrapper[4960]: I1002 07:50:44.069181 4960 scope.go:117] "RemoveContainer" containerID="f6b3d4d3ac0772db8208b0643fc13ec6b76610dfa4a9c4d64e5625c1843f9d40" Oct 02 07:50:44 crc kubenswrapper[4960]: I1002 07:50:44.134312 4960 scope.go:117] "RemoveContainer" containerID="c0b040b5f28ce1989e16094bd189d8de6a543e5af0f71f799568e35acad5be5e" Oct 02 07:50:44 crc kubenswrapper[4960]: I1002 07:50:44.215153 4960 scope.go:117] "RemoveContainer" containerID="7f3e3bf87367fa4cc7d2bb7d03f02691e5bb4878a2954f8c4844e82397b7143e" Oct 02 07:50:44 crc kubenswrapper[4960]: I1002 07:50:44.250606 4960 scope.go:117] "RemoveContainer" containerID="58b10688097179490c642ef7ff2e726b40b19eeca457354427a375a4536d0392" Oct 02 07:50:44 crc kubenswrapper[4960]: I1002 07:50:44.312623 4960 scope.go:117] "RemoveContainer" containerID="72151a5c2fc6d9a50f9f5eb1970d291b79db641639cb3d48ff61df9d042eb246" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.286265 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r9grs"] Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.290731 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.300360 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcxxf\" (UniqueName: \"kubernetes.io/projected/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-kube-api-access-wcxxf\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.300452 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-utilities\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.300626 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-catalog-content\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.310908 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r9grs"] Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.403732 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-catalog-content\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.404113 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcxxf\" (UniqueName: \"kubernetes.io/projected/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-kube-api-access-wcxxf\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.404174 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-utilities\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.404873 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-catalog-content\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.405496 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-utilities\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.447153 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcxxf\" (UniqueName: \"kubernetes.io/projected/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-kube-api-access-wcxxf\") pod \"certified-operators-r9grs\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:08 crc kubenswrapper[4960]: I1002 07:51:08.619940 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:09 crc kubenswrapper[4960]: I1002 07:51:09.133719 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r9grs"] Oct 02 07:51:09 crc kubenswrapper[4960]: I1002 07:51:09.371161 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9grs" event={"ID":"ccc4f083-178f-416d-86b4-e8d07a1ff3d3","Type":"ContainerStarted","Data":"991057d3fda1562fab0bdee53abc15ec184257cb8f7ea3ce4b1e0205248cbffb"} Oct 02 07:51:10 crc kubenswrapper[4960]: I1002 07:51:10.385050 4960 generic.go:334] "Generic (PLEG): container finished" podID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerID="066f6ef80720dba26cb30f03f8c96b3d77166e7ae3634422f568c3a973f9b72c" exitCode=0 Oct 02 07:51:10 crc kubenswrapper[4960]: I1002 07:51:10.385595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9grs" event={"ID":"ccc4f083-178f-416d-86b4-e8d07a1ff3d3","Type":"ContainerDied","Data":"066f6ef80720dba26cb30f03f8c96b3d77166e7ae3634422f568c3a973f9b72c"} Oct 02 07:51:11 crc kubenswrapper[4960]: I1002 07:51:11.398205 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9grs" event={"ID":"ccc4f083-178f-416d-86b4-e8d07a1ff3d3","Type":"ContainerStarted","Data":"766e38df662665e37ed10a9ec73e37bf8bea511d77512a034f24a32d8bbc0f5d"} Oct 02 07:51:12 crc kubenswrapper[4960]: I1002 07:51:12.407955 4960 generic.go:334] "Generic (PLEG): container finished" podID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerID="766e38df662665e37ed10a9ec73e37bf8bea511d77512a034f24a32d8bbc0f5d" exitCode=0 Oct 02 07:51:12 crc kubenswrapper[4960]: I1002 07:51:12.408078 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9grs" event={"ID":"ccc4f083-178f-416d-86b4-e8d07a1ff3d3","Type":"ContainerDied","Data":"766e38df662665e37ed10a9ec73e37bf8bea511d77512a034f24a32d8bbc0f5d"} Oct 02 07:51:13 crc kubenswrapper[4960]: I1002 07:51:13.419300 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9grs" event={"ID":"ccc4f083-178f-416d-86b4-e8d07a1ff3d3","Type":"ContainerStarted","Data":"c5112ee4291c64553463a2037fb8377c97318619afa4aad9f981071d63f0bdb9"} Oct 02 07:51:13 crc kubenswrapper[4960]: I1002 07:51:13.443601 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r9grs" podStartSLOduration=3.012324438 podStartE2EDuration="5.443577061s" podCreationTimestamp="2025-10-02 07:51:08 +0000 UTC" firstStartedPulling="2025-10-02 07:51:10.387675355 +0000 UTC m=+2091.419621642" lastFinishedPulling="2025-10-02 07:51:12.818927978 +0000 UTC m=+2093.850874265" observedRunningTime="2025-10-02 07:51:13.438547681 +0000 UTC m=+2094.470493978" watchObservedRunningTime="2025-10-02 07:51:13.443577061 +0000 UTC m=+2094.475523348" Oct 02 07:51:18 crc kubenswrapper[4960]: I1002 07:51:18.620672 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:18 crc kubenswrapper[4960]: I1002 07:51:18.621257 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:18 crc kubenswrapper[4960]: I1002 07:51:18.664824 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:19 crc kubenswrapper[4960]: I1002 07:51:19.541474 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:19 crc kubenswrapper[4960]: I1002 07:51:19.599295 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r9grs"] Oct 02 07:51:21 crc kubenswrapper[4960]: I1002 07:51:21.509626 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r9grs" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="registry-server" containerID="cri-o://c5112ee4291c64553463a2037fb8377c97318619afa4aad9f981071d63f0bdb9" gracePeriod=2 Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.521254 4960 generic.go:334] "Generic (PLEG): container finished" podID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerID="c5112ee4291c64553463a2037fb8377c97318619afa4aad9f981071d63f0bdb9" exitCode=0 Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.521464 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9grs" event={"ID":"ccc4f083-178f-416d-86b4-e8d07a1ff3d3","Type":"ContainerDied","Data":"c5112ee4291c64553463a2037fb8377c97318619afa4aad9f981071d63f0bdb9"} Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.521756 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9grs" event={"ID":"ccc4f083-178f-416d-86b4-e8d07a1ff3d3","Type":"ContainerDied","Data":"991057d3fda1562fab0bdee53abc15ec184257cb8f7ea3ce4b1e0205248cbffb"} Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.521795 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="991057d3fda1562fab0bdee53abc15ec184257cb8f7ea3ce4b1e0205248cbffb" Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.554990 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.702905 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcxxf\" (UniqueName: \"kubernetes.io/projected/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-kube-api-access-wcxxf\") pod \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.703417 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-utilities\") pod \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.703689 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-catalog-content\") pod \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\" (UID: \"ccc4f083-178f-416d-86b4-e8d07a1ff3d3\") " Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.704424 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-utilities" (OuterVolumeSpecName: "utilities") pod "ccc4f083-178f-416d-86b4-e8d07a1ff3d3" (UID: "ccc4f083-178f-416d-86b4-e8d07a1ff3d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.704934 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.715426 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-kube-api-access-wcxxf" (OuterVolumeSpecName: "kube-api-access-wcxxf") pod "ccc4f083-178f-416d-86b4-e8d07a1ff3d3" (UID: "ccc4f083-178f-416d-86b4-e8d07a1ff3d3"). InnerVolumeSpecName "kube-api-access-wcxxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.783214 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccc4f083-178f-416d-86b4-e8d07a1ff3d3" (UID: "ccc4f083-178f-416d-86b4-e8d07a1ff3d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.808314 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcxxf\" (UniqueName: \"kubernetes.io/projected/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-kube-api-access-wcxxf\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:22 crc kubenswrapper[4960]: I1002 07:51:22.808400 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc4f083-178f-416d-86b4-e8d07a1ff3d3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:23 crc kubenswrapper[4960]: I1002 07:51:23.535428 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9grs" Oct 02 07:51:23 crc kubenswrapper[4960]: I1002 07:51:23.589836 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r9grs"] Oct 02 07:51:23 crc kubenswrapper[4960]: I1002 07:51:23.598858 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r9grs"] Oct 02 07:51:24 crc kubenswrapper[4960]: I1002 07:51:24.355091 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" path="/var/lib/kubelet/pods/ccc4f083-178f-416d-86b4-e8d07a1ff3d3/volumes" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.672736 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7fg8p"] Oct 02 07:51:32 crc kubenswrapper[4960]: E1002 07:51:32.674492 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="extract-content" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.674507 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="extract-content" Oct 02 07:51:32 crc kubenswrapper[4960]: E1002 07:51:32.674519 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="registry-server" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.674524 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="registry-server" Oct 02 07:51:32 crc kubenswrapper[4960]: E1002 07:51:32.674538 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="extract-utilities" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.674544 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="extract-utilities" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.674806 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc4f083-178f-416d-86b4-e8d07a1ff3d3" containerName="registry-server" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.676539 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.740908 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fg8p"] Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.856135 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-utilities\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.856188 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-catalog-content\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.856695 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb2dz\" (UniqueName: \"kubernetes.io/projected/cfc68c94-d63d-43d1-a18b-0de6e9739439-kube-api-access-tb2dz\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.959914 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-utilities\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.960108 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-catalog-content\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.960356 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb2dz\" (UniqueName: \"kubernetes.io/projected/cfc68c94-d63d-43d1-a18b-0de6e9739439-kube-api-access-tb2dz\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.960829 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-utilities\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.960844 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-catalog-content\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:32 crc kubenswrapper[4960]: I1002 07:51:32.982239 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb2dz\" (UniqueName: \"kubernetes.io/projected/cfc68c94-d63d-43d1-a18b-0de6e9739439-kube-api-access-tb2dz\") pod \"community-operators-7fg8p\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:33 crc kubenswrapper[4960]: I1002 07:51:33.060909 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:33 crc kubenswrapper[4960]: I1002 07:51:33.593925 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fg8p"] Oct 02 07:51:33 crc kubenswrapper[4960]: I1002 07:51:33.670610 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fg8p" event={"ID":"cfc68c94-d63d-43d1-a18b-0de6e9739439","Type":"ContainerStarted","Data":"47ab6414c7659b468ed210cc6ee6cdc7cfea3fb9169976341ee9862f1a3bf31d"} Oct 02 07:51:34 crc kubenswrapper[4960]: I1002 07:51:34.685095 4960 generic.go:334] "Generic (PLEG): container finished" podID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerID="0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039" exitCode=0 Oct 02 07:51:34 crc kubenswrapper[4960]: I1002 07:51:34.685190 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fg8p" event={"ID":"cfc68c94-d63d-43d1-a18b-0de6e9739439","Type":"ContainerDied","Data":"0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039"} Oct 02 07:51:35 crc kubenswrapper[4960]: I1002 07:51:35.698726 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fg8p" event={"ID":"cfc68c94-d63d-43d1-a18b-0de6e9739439","Type":"ContainerStarted","Data":"86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47"} Oct 02 07:51:36 crc kubenswrapper[4960]: I1002 07:51:36.713125 4960 generic.go:334] "Generic (PLEG): container finished" podID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerID="86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47" exitCode=0 Oct 02 07:51:36 crc kubenswrapper[4960]: I1002 07:51:36.713199 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fg8p" event={"ID":"cfc68c94-d63d-43d1-a18b-0de6e9739439","Type":"ContainerDied","Data":"86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47"} Oct 02 07:51:37 crc kubenswrapper[4960]: I1002 07:51:37.726321 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fg8p" event={"ID":"cfc68c94-d63d-43d1-a18b-0de6e9739439","Type":"ContainerStarted","Data":"b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13"} Oct 02 07:51:37 crc kubenswrapper[4960]: I1002 07:51:37.756175 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7fg8p" podStartSLOduration=3.142265773 podStartE2EDuration="5.75615133s" podCreationTimestamp="2025-10-02 07:51:32 +0000 UTC" firstStartedPulling="2025-10-02 07:51:34.687950687 +0000 UTC m=+2115.719896984" lastFinishedPulling="2025-10-02 07:51:37.301836254 +0000 UTC m=+2118.333782541" observedRunningTime="2025-10-02 07:51:37.752041014 +0000 UTC m=+2118.783987311" watchObservedRunningTime="2025-10-02 07:51:37.75615133 +0000 UTC m=+2118.788097617" Oct 02 07:51:43 crc kubenswrapper[4960]: I1002 07:51:43.061463 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:43 crc kubenswrapper[4960]: I1002 07:51:43.062254 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:43 crc kubenswrapper[4960]: I1002 07:51:43.135955 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:43 crc kubenswrapper[4960]: I1002 07:51:43.883162 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:43 crc kubenswrapper[4960]: I1002 07:51:43.939928 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fg8p"] Oct 02 07:51:44 crc kubenswrapper[4960]: I1002 07:51:44.494761 4960 scope.go:117] "RemoveContainer" containerID="2bd9cb016a0bd6706ec701b199c65be55869d511eed34e54b9a3f245df73d579" Oct 02 07:51:44 crc kubenswrapper[4960]: I1002 07:51:44.537762 4960 scope.go:117] "RemoveContainer" containerID="447c18823fb4d0ce64a5acf74211aaa51b2c00eca2b9c57a74534e00265cbca4" Oct 02 07:51:44 crc kubenswrapper[4960]: I1002 07:51:44.592579 4960 scope.go:117] "RemoveContainer" containerID="0795d5125ef89562e2e96210224e40f0cc4271e33a229b9ad06d52c9c206acb5" Oct 02 07:51:45 crc kubenswrapper[4960]: I1002 07:51:45.841281 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7fg8p" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="registry-server" containerID="cri-o://b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13" gracePeriod=2 Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.313035 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.492812 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-utilities\") pod \"cfc68c94-d63d-43d1-a18b-0de6e9739439\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.493035 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb2dz\" (UniqueName: \"kubernetes.io/projected/cfc68c94-d63d-43d1-a18b-0de6e9739439-kube-api-access-tb2dz\") pod \"cfc68c94-d63d-43d1-a18b-0de6e9739439\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.493193 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-catalog-content\") pod \"cfc68c94-d63d-43d1-a18b-0de6e9739439\" (UID: \"cfc68c94-d63d-43d1-a18b-0de6e9739439\") " Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.494578 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-utilities" (OuterVolumeSpecName: "utilities") pod "cfc68c94-d63d-43d1-a18b-0de6e9739439" (UID: "cfc68c94-d63d-43d1-a18b-0de6e9739439"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.504549 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc68c94-d63d-43d1-a18b-0de6e9739439-kube-api-access-tb2dz" (OuterVolumeSpecName: "kube-api-access-tb2dz") pod "cfc68c94-d63d-43d1-a18b-0de6e9739439" (UID: "cfc68c94-d63d-43d1-a18b-0de6e9739439"). InnerVolumeSpecName "kube-api-access-tb2dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.554159 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfc68c94-d63d-43d1-a18b-0de6e9739439" (UID: "cfc68c94-d63d-43d1-a18b-0de6e9739439"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.595942 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.596361 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb2dz\" (UniqueName: \"kubernetes.io/projected/cfc68c94-d63d-43d1-a18b-0de6e9739439-kube-api-access-tb2dz\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.596442 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc68c94-d63d-43d1-a18b-0de6e9739439-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.853462 4960 generic.go:334] "Generic (PLEG): container finished" podID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerID="b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13" exitCode=0 Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.853515 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fg8p" event={"ID":"cfc68c94-d63d-43d1-a18b-0de6e9739439","Type":"ContainerDied","Data":"b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13"} Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.853549 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fg8p" event={"ID":"cfc68c94-d63d-43d1-a18b-0de6e9739439","Type":"ContainerDied","Data":"47ab6414c7659b468ed210cc6ee6cdc7cfea3fb9169976341ee9862f1a3bf31d"} Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.853571 4960 scope.go:117] "RemoveContainer" containerID="b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.853626 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fg8p" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.878932 4960 scope.go:117] "RemoveContainer" containerID="86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.915061 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fg8p"] Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.921933 4960 scope.go:117] "RemoveContainer" containerID="0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.925164 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7fg8p"] Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.943234 4960 scope.go:117] "RemoveContainer" containerID="b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13" Oct 02 07:51:46 crc kubenswrapper[4960]: E1002 07:51:46.943768 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13\": container with ID starting with b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13 not found: ID does not exist" containerID="b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.943827 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13"} err="failed to get container status \"b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13\": rpc error: code = NotFound desc = could not find container \"b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13\": container with ID starting with b3ecc3c690f088a0fcba668f60dac1c55a1d37b2f05613f4d05eebdf35fb3d13 not found: ID does not exist" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.943856 4960 scope.go:117] "RemoveContainer" containerID="86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47" Oct 02 07:51:46 crc kubenswrapper[4960]: E1002 07:51:46.944368 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47\": container with ID starting with 86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47 not found: ID does not exist" containerID="86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.944432 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47"} err="failed to get container status \"86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47\": rpc error: code = NotFound desc = could not find container \"86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47\": container with ID starting with 86839374ccfb21ffc25b58a9f356db1783abf2e482062929ab3407d538529e47 not found: ID does not exist" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.944470 4960 scope.go:117] "RemoveContainer" containerID="0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039" Oct 02 07:51:46 crc kubenswrapper[4960]: E1002 07:51:46.944829 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039\": container with ID starting with 0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039 not found: ID does not exist" containerID="0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039" Oct 02 07:51:46 crc kubenswrapper[4960]: I1002 07:51:46.944853 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039"} err="failed to get container status \"0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039\": rpc error: code = NotFound desc = could not find container \"0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039\": container with ID starting with 0237ed4d1246550ad9e5c81105063fd2d43aa0d936913084a9ce9d0d94a75039 not found: ID does not exist" Oct 02 07:51:48 crc kubenswrapper[4960]: I1002 07:51:48.342240 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" path="/var/lib/kubelet/pods/cfc68c94-d63d-43d1-a18b-0de6e9739439/volumes" Oct 02 07:52:17 crc kubenswrapper[4960]: I1002 07:52:17.160659 4960 generic.go:334] "Generic (PLEG): container finished" podID="9d905943-1348-47e8-81a5-5794bb84f7ff" containerID="4ad048516b2480ea54f8806270a893d7907b7280e086eb597f2e389b416fa3c3" exitCode=0 Oct 02 07:52:17 crc kubenswrapper[4960]: I1002 07:52:17.160817 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" event={"ID":"9d905943-1348-47e8-81a5-5794bb84f7ff","Type":"ContainerDied","Data":"4ad048516b2480ea54f8806270a893d7907b7280e086eb597f2e389b416fa3c3"} Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.677467 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.828631 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ceph\") pod \"9d905943-1348-47e8-81a5-5794bb84f7ff\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.828806 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2s6z\" (UniqueName: \"kubernetes.io/projected/9d905943-1348-47e8-81a5-5794bb84f7ff-kube-api-access-b2s6z\") pod \"9d905943-1348-47e8-81a5-5794bb84f7ff\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.828835 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-inventory\") pod \"9d905943-1348-47e8-81a5-5794bb84f7ff\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.828948 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ssh-key\") pod \"9d905943-1348-47e8-81a5-5794bb84f7ff\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.829077 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-bootstrap-combined-ca-bundle\") pod \"9d905943-1348-47e8-81a5-5794bb84f7ff\" (UID: \"9d905943-1348-47e8-81a5-5794bb84f7ff\") " Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.837288 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ceph" (OuterVolumeSpecName: "ceph") pod "9d905943-1348-47e8-81a5-5794bb84f7ff" (UID: "9d905943-1348-47e8-81a5-5794bb84f7ff"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.843208 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d905943-1348-47e8-81a5-5794bb84f7ff-kube-api-access-b2s6z" (OuterVolumeSpecName: "kube-api-access-b2s6z") pod "9d905943-1348-47e8-81a5-5794bb84f7ff" (UID: "9d905943-1348-47e8-81a5-5794bb84f7ff"). InnerVolumeSpecName "kube-api-access-b2s6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.854246 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9d905943-1348-47e8-81a5-5794bb84f7ff" (UID: "9d905943-1348-47e8-81a5-5794bb84f7ff"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.866085 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d905943-1348-47e8-81a5-5794bb84f7ff" (UID: "9d905943-1348-47e8-81a5-5794bb84f7ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.885649 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-inventory" (OuterVolumeSpecName: "inventory") pod "9d905943-1348-47e8-81a5-5794bb84f7ff" (UID: "9d905943-1348-47e8-81a5-5794bb84f7ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.931636 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2s6z\" (UniqueName: \"kubernetes.io/projected/9d905943-1348-47e8-81a5-5794bb84f7ff-kube-api-access-b2s6z\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.931673 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.931686 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.931699 4960 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:18 crc kubenswrapper[4960]: I1002 07:52:18.931714 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9d905943-1348-47e8-81a5-5794bb84f7ff-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.182424 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" event={"ID":"9d905943-1348-47e8-81a5-5794bb84f7ff","Type":"ContainerDied","Data":"e776d0605b6f612bd9f1a46dabae5611a6d53b5befdf1b53d96698f94961b456"} Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.182474 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e776d0605b6f612bd9f1a46dabae5611a6d53b5befdf1b53d96698f94961b456" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.182486 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.288301 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt"] Oct 02 07:52:19 crc kubenswrapper[4960]: E1002 07:52:19.288766 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="extract-content" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.288789 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="extract-content" Oct 02 07:52:19 crc kubenswrapper[4960]: E1002 07:52:19.288819 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="registry-server" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.288828 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="registry-server" Oct 02 07:52:19 crc kubenswrapper[4960]: E1002 07:52:19.288836 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d905943-1348-47e8-81a5-5794bb84f7ff" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.288846 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d905943-1348-47e8-81a5-5794bb84f7ff" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:19 crc kubenswrapper[4960]: E1002 07:52:19.288861 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="extract-utilities" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.288870 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="extract-utilities" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.289131 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc68c94-d63d-43d1-a18b-0de6e9739439" containerName="registry-server" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.289164 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d905943-1348-47e8-81a5-5794bb84f7ff" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.289996 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.292414 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.293188 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.294040 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.294421 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.296394 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.309608 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt"] Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.450818 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.451233 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.451335 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczj8\" (UniqueName: \"kubernetes.io/projected/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-kube-api-access-pczj8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.451664 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.553959 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.554046 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.554088 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczj8\" (UniqueName: \"kubernetes.io/projected/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-kube-api-access-pczj8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.554138 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.559280 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.559511 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.563508 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.580459 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczj8\" (UniqueName: \"kubernetes.io/projected/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-kube-api-access-pczj8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:19 crc kubenswrapper[4960]: I1002 07:52:19.648040 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.082723 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sr899"] Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.084764 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.103726 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sr899"] Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.169343 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-utilities\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.169492 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxvgt\" (UniqueName: \"kubernetes.io/projected/b51220e7-4162-4ecb-a06f-3fe6330de5c9-kube-api-access-sxvgt\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.169584 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-catalog-content\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.243424 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt"] Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.271821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxvgt\" (UniqueName: \"kubernetes.io/projected/b51220e7-4162-4ecb-a06f-3fe6330de5c9-kube-api-access-sxvgt\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.271940 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-catalog-content\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.272090 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-utilities\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.272833 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-utilities\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.273612 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-catalog-content\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.292513 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxvgt\" (UniqueName: \"kubernetes.io/projected/b51220e7-4162-4ecb-a06f-3fe6330de5c9-kube-api-access-sxvgt\") pod \"redhat-operators-sr899\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.412886 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.756209 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:52:20 crc kubenswrapper[4960]: I1002 07:52:20.914813 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sr899"] Oct 02 07:52:21 crc kubenswrapper[4960]: I1002 07:52:21.202822 4960 generic.go:334] "Generic (PLEG): container finished" podID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerID="ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca" exitCode=0 Oct 02 07:52:21 crc kubenswrapper[4960]: I1002 07:52:21.202946 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sr899" event={"ID":"b51220e7-4162-4ecb-a06f-3fe6330de5c9","Type":"ContainerDied","Data":"ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca"} Oct 02 07:52:21 crc kubenswrapper[4960]: I1002 07:52:21.203057 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sr899" event={"ID":"b51220e7-4162-4ecb-a06f-3fe6330de5c9","Type":"ContainerStarted","Data":"bbcfbb5dec577f9ced53b85fcad7926e30848ba59baffc13ed9961238c6c510d"} Oct 02 07:52:21 crc kubenswrapper[4960]: I1002 07:52:21.205273 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" event={"ID":"947ad109-ebe1-4de0-ad2e-4fa4ae75b238","Type":"ContainerStarted","Data":"cc672045f9e6be5b68a79ef93ceff916d5dcac826c0774b91ebba48b0ed265f1"} Oct 02 07:52:21 crc kubenswrapper[4960]: I1002 07:52:21.205301 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" event={"ID":"947ad109-ebe1-4de0-ad2e-4fa4ae75b238","Type":"ContainerStarted","Data":"9b8ac9eb565003b11fd53d5f73a499a21649000dfd289ee688b6beef017b565b"} Oct 02 07:52:21 crc kubenswrapper[4960]: I1002 07:52:21.257563 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" podStartSLOduration=1.763476482 podStartE2EDuration="2.2575356s" podCreationTimestamp="2025-10-02 07:52:19 +0000 UTC" firstStartedPulling="2025-10-02 07:52:20.259136223 +0000 UTC m=+2161.291082510" lastFinishedPulling="2025-10-02 07:52:20.753195341 +0000 UTC m=+2161.785141628" observedRunningTime="2025-10-02 07:52:21.254899541 +0000 UTC m=+2162.286845828" watchObservedRunningTime="2025-10-02 07:52:21.2575356 +0000 UTC m=+2162.289481887" Oct 02 07:52:23 crc kubenswrapper[4960]: I1002 07:52:23.241586 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sr899" event={"ID":"b51220e7-4162-4ecb-a06f-3fe6330de5c9","Type":"ContainerStarted","Data":"b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab"} Oct 02 07:52:24 crc kubenswrapper[4960]: I1002 07:52:24.258578 4960 generic.go:334] "Generic (PLEG): container finished" podID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerID="b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab" exitCode=0 Oct 02 07:52:24 crc kubenswrapper[4960]: I1002 07:52:24.258685 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sr899" event={"ID":"b51220e7-4162-4ecb-a06f-3fe6330de5c9","Type":"ContainerDied","Data":"b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab"} Oct 02 07:52:25 crc kubenswrapper[4960]: I1002 07:52:25.271461 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sr899" event={"ID":"b51220e7-4162-4ecb-a06f-3fe6330de5c9","Type":"ContainerStarted","Data":"ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b"} Oct 02 07:52:25 crc kubenswrapper[4960]: I1002 07:52:25.298567 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sr899" podStartSLOduration=1.805823765 podStartE2EDuration="5.298532288s" podCreationTimestamp="2025-10-02 07:52:20 +0000 UTC" firstStartedPulling="2025-10-02 07:52:21.205208955 +0000 UTC m=+2162.237155242" lastFinishedPulling="2025-10-02 07:52:24.697917478 +0000 UTC m=+2165.729863765" observedRunningTime="2025-10-02 07:52:25.293867215 +0000 UTC m=+2166.325813502" watchObservedRunningTime="2025-10-02 07:52:25.298532288 +0000 UTC m=+2166.330478615" Oct 02 07:52:29 crc kubenswrapper[4960]: I1002 07:52:29.150166 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:52:29 crc kubenswrapper[4960]: I1002 07:52:29.151392 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:52:30 crc kubenswrapper[4960]: I1002 07:52:30.414219 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:30 crc kubenswrapper[4960]: I1002 07:52:30.414303 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:30 crc kubenswrapper[4960]: I1002 07:52:30.468166 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:31 crc kubenswrapper[4960]: I1002 07:52:31.378134 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:31 crc kubenswrapper[4960]: I1002 07:52:31.453535 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sr899"] Oct 02 07:52:33 crc kubenswrapper[4960]: I1002 07:52:33.352704 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sr899" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="registry-server" containerID="cri-o://ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b" gracePeriod=2 Oct 02 07:52:33 crc kubenswrapper[4960]: I1002 07:52:33.911150 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:33 crc kubenswrapper[4960]: I1002 07:52:33.994932 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-catalog-content\") pod \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " Oct 02 07:52:33 crc kubenswrapper[4960]: I1002 07:52:33.995158 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxvgt\" (UniqueName: \"kubernetes.io/projected/b51220e7-4162-4ecb-a06f-3fe6330de5c9-kube-api-access-sxvgt\") pod \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " Oct 02 07:52:33 crc kubenswrapper[4960]: I1002 07:52:33.995248 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-utilities\") pod \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\" (UID: \"b51220e7-4162-4ecb-a06f-3fe6330de5c9\") " Oct 02 07:52:33 crc kubenswrapper[4960]: I1002 07:52:33.996054 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-utilities" (OuterVolumeSpecName: "utilities") pod "b51220e7-4162-4ecb-a06f-3fe6330de5c9" (UID: "b51220e7-4162-4ecb-a06f-3fe6330de5c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.002621 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51220e7-4162-4ecb-a06f-3fe6330de5c9-kube-api-access-sxvgt" (OuterVolumeSpecName: "kube-api-access-sxvgt") pod "b51220e7-4162-4ecb-a06f-3fe6330de5c9" (UID: "b51220e7-4162-4ecb-a06f-3fe6330de5c9"). InnerVolumeSpecName "kube-api-access-sxvgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.097595 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxvgt\" (UniqueName: \"kubernetes.io/projected/b51220e7-4162-4ecb-a06f-3fe6330de5c9-kube-api-access-sxvgt\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.097643 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.111267 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b51220e7-4162-4ecb-a06f-3fe6330de5c9" (UID: "b51220e7-4162-4ecb-a06f-3fe6330de5c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.200465 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b51220e7-4162-4ecb-a06f-3fe6330de5c9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.364779 4960 generic.go:334] "Generic (PLEG): container finished" podID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerID="ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b" exitCode=0 Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.364843 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sr899" event={"ID":"b51220e7-4162-4ecb-a06f-3fe6330de5c9","Type":"ContainerDied","Data":"ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b"} Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.364882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sr899" event={"ID":"b51220e7-4162-4ecb-a06f-3fe6330de5c9","Type":"ContainerDied","Data":"bbcfbb5dec577f9ced53b85fcad7926e30848ba59baffc13ed9961238c6c510d"} Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.364906 4960 scope.go:117] "RemoveContainer" containerID="ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.365281 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sr899" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.401815 4960 scope.go:117] "RemoveContainer" containerID="b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.410571 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sr899"] Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.434745 4960 scope.go:117] "RemoveContainer" containerID="ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.434970 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sr899"] Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.492271 4960 scope.go:117] "RemoveContainer" containerID="ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b" Oct 02 07:52:34 crc kubenswrapper[4960]: E1002 07:52:34.493286 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b\": container with ID starting with ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b not found: ID does not exist" containerID="ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.493370 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b"} err="failed to get container status \"ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b\": rpc error: code = NotFound desc = could not find container \"ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b\": container with ID starting with ce41ac17567d2897449b58dd6597dcfb051307ed57d9d3a8112c547869ec2e0b not found: ID does not exist" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.493434 4960 scope.go:117] "RemoveContainer" containerID="b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab" Oct 02 07:52:34 crc kubenswrapper[4960]: E1002 07:52:34.494219 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab\": container with ID starting with b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab not found: ID does not exist" containerID="b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.494315 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab"} err="failed to get container status \"b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab\": rpc error: code = NotFound desc = could not find container \"b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab\": container with ID starting with b56fca29d0d7e5aa6c1d85ec45fdd41016dd8f5c8364061ce0d2fa7fa158e5ab not found: ID does not exist" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.494371 4960 scope.go:117] "RemoveContainer" containerID="ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca" Oct 02 07:52:34 crc kubenswrapper[4960]: E1002 07:52:34.495072 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca\": container with ID starting with ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca not found: ID does not exist" containerID="ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca" Oct 02 07:52:34 crc kubenswrapper[4960]: I1002 07:52:34.495116 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca"} err="failed to get container status \"ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca\": rpc error: code = NotFound desc = could not find container \"ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca\": container with ID starting with ab959c04371fb13d4a2165e876c02932500799717d31e8793c32d733156718ca not found: ID does not exist" Oct 02 07:52:36 crc kubenswrapper[4960]: I1002 07:52:36.345933 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" path="/var/lib/kubelet/pods/b51220e7-4162-4ecb-a06f-3fe6330de5c9/volumes" Oct 02 07:52:46 crc kubenswrapper[4960]: I1002 07:52:46.472403 4960 generic.go:334] "Generic (PLEG): container finished" podID="947ad109-ebe1-4de0-ad2e-4fa4ae75b238" containerID="cc672045f9e6be5b68a79ef93ceff916d5dcac826c0774b91ebba48b0ed265f1" exitCode=0 Oct 02 07:52:46 crc kubenswrapper[4960]: I1002 07:52:46.472494 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" event={"ID":"947ad109-ebe1-4de0-ad2e-4fa4ae75b238","Type":"ContainerDied","Data":"cc672045f9e6be5b68a79ef93ceff916d5dcac826c0774b91ebba48b0ed265f1"} Oct 02 07:52:47 crc kubenswrapper[4960]: I1002 07:52:47.928787 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.009925 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pczj8\" (UniqueName: \"kubernetes.io/projected/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-kube-api-access-pczj8\") pod \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.010149 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-inventory\") pod \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.010194 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ssh-key\") pod \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.010277 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ceph\") pod \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\" (UID: \"947ad109-ebe1-4de0-ad2e-4fa4ae75b238\") " Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.020954 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ceph" (OuterVolumeSpecName: "ceph") pod "947ad109-ebe1-4de0-ad2e-4fa4ae75b238" (UID: "947ad109-ebe1-4de0-ad2e-4fa4ae75b238"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.021789 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-kube-api-access-pczj8" (OuterVolumeSpecName: "kube-api-access-pczj8") pod "947ad109-ebe1-4de0-ad2e-4fa4ae75b238" (UID: "947ad109-ebe1-4de0-ad2e-4fa4ae75b238"). InnerVolumeSpecName "kube-api-access-pczj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.040010 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "947ad109-ebe1-4de0-ad2e-4fa4ae75b238" (UID: "947ad109-ebe1-4de0-ad2e-4fa4ae75b238"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.041699 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-inventory" (OuterVolumeSpecName: "inventory") pod "947ad109-ebe1-4de0-ad2e-4fa4ae75b238" (UID: "947ad109-ebe1-4de0-ad2e-4fa4ae75b238"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.112201 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.112237 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.112248 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.112257 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pczj8\" (UniqueName: \"kubernetes.io/projected/947ad109-ebe1-4de0-ad2e-4fa4ae75b238-kube-api-access-pczj8\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.491025 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" event={"ID":"947ad109-ebe1-4de0-ad2e-4fa4ae75b238","Type":"ContainerDied","Data":"9b8ac9eb565003b11fd53d5f73a499a21649000dfd289ee688b6beef017b565b"} Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.491076 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b8ac9eb565003b11fd53d5f73a499a21649000dfd289ee688b6beef017b565b" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.491198 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.587653 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j"] Oct 02 07:52:48 crc kubenswrapper[4960]: E1002 07:52:48.588485 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="registry-server" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.588560 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="registry-server" Oct 02 07:52:48 crc kubenswrapper[4960]: E1002 07:52:48.588677 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="extract-content" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.588764 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="extract-content" Oct 02 07:52:48 crc kubenswrapper[4960]: E1002 07:52:48.588865 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947ad109-ebe1-4de0-ad2e-4fa4ae75b238" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.588947 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="947ad109-ebe1-4de0-ad2e-4fa4ae75b238" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:48 crc kubenswrapper[4960]: E1002 07:52:48.589057 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="extract-utilities" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.589121 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="extract-utilities" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.589377 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="947ad109-ebe1-4de0-ad2e-4fa4ae75b238" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.589495 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51220e7-4162-4ecb-a06f-3fe6330de5c9" containerName="registry-server" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.590343 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.596476 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.596483 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.596738 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.596926 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.597135 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.607914 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j"] Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.730002 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.730076 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzn8l\" (UniqueName: \"kubernetes.io/projected/5488228f-8fc0-46cb-be89-51d65f7ccff8-kube-api-access-jzn8l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.730226 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.730514 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.833349 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.834064 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzn8l\" (UniqueName: \"kubernetes.io/projected/5488228f-8fc0-46cb-be89-51d65f7ccff8-kube-api-access-jzn8l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.834164 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.834260 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.841904 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.843023 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.846627 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.852710 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzn8l\" (UniqueName: \"kubernetes.io/projected/5488228f-8fc0-46cb-be89-51d65f7ccff8-kube-api-access-jzn8l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:48 crc kubenswrapper[4960]: I1002 07:52:48.915745 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:49 crc kubenswrapper[4960]: I1002 07:52:49.514312 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j"] Oct 02 07:52:50 crc kubenswrapper[4960]: I1002 07:52:50.511163 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" event={"ID":"5488228f-8fc0-46cb-be89-51d65f7ccff8","Type":"ContainerStarted","Data":"7e0bee01e19cc0ce5955626c824023e2dc841e8d5431e1435371c280f1a8b2c7"} Oct 02 07:52:50 crc kubenswrapper[4960]: I1002 07:52:50.511594 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" event={"ID":"5488228f-8fc0-46cb-be89-51d65f7ccff8","Type":"ContainerStarted","Data":"c88da6d7deb3f7d418f7af615e7752c5fca795e48f54f34c856119cddb0ce50e"} Oct 02 07:52:50 crc kubenswrapper[4960]: I1002 07:52:50.533396 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" podStartSLOduration=1.9718967059999999 podStartE2EDuration="2.533375637s" podCreationTimestamp="2025-10-02 07:52:48 +0000 UTC" firstStartedPulling="2025-10-02 07:52:49.52317653 +0000 UTC m=+2190.555122807" lastFinishedPulling="2025-10-02 07:52:50.084655461 +0000 UTC m=+2191.116601738" observedRunningTime="2025-10-02 07:52:50.53231069 +0000 UTC m=+2191.564256987" watchObservedRunningTime="2025-10-02 07:52:50.533375637 +0000 UTC m=+2191.565321934" Oct 02 07:52:55 crc kubenswrapper[4960]: I1002 07:52:55.560994 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" event={"ID":"5488228f-8fc0-46cb-be89-51d65f7ccff8","Type":"ContainerDied","Data":"7e0bee01e19cc0ce5955626c824023e2dc841e8d5431e1435371c280f1a8b2c7"} Oct 02 07:52:55 crc kubenswrapper[4960]: I1002 07:52:55.560868 4960 generic.go:334] "Generic (PLEG): container finished" podID="5488228f-8fc0-46cb-be89-51d65f7ccff8" containerID="7e0bee01e19cc0ce5955626c824023e2dc841e8d5431e1435371c280f1a8b2c7" exitCode=0 Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.101598 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.233519 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ssh-key\") pod \"5488228f-8fc0-46cb-be89-51d65f7ccff8\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.233595 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ceph\") pod \"5488228f-8fc0-46cb-be89-51d65f7ccff8\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.233713 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzn8l\" (UniqueName: \"kubernetes.io/projected/5488228f-8fc0-46cb-be89-51d65f7ccff8-kube-api-access-jzn8l\") pod \"5488228f-8fc0-46cb-be89-51d65f7ccff8\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.233814 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-inventory\") pod \"5488228f-8fc0-46cb-be89-51d65f7ccff8\" (UID: \"5488228f-8fc0-46cb-be89-51d65f7ccff8\") " Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.242059 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ceph" (OuterVolumeSpecName: "ceph") pod "5488228f-8fc0-46cb-be89-51d65f7ccff8" (UID: "5488228f-8fc0-46cb-be89-51d65f7ccff8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.256171 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5488228f-8fc0-46cb-be89-51d65f7ccff8-kube-api-access-jzn8l" (OuterVolumeSpecName: "kube-api-access-jzn8l") pod "5488228f-8fc0-46cb-be89-51d65f7ccff8" (UID: "5488228f-8fc0-46cb-be89-51d65f7ccff8"). InnerVolumeSpecName "kube-api-access-jzn8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.271767 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5488228f-8fc0-46cb-be89-51d65f7ccff8" (UID: "5488228f-8fc0-46cb-be89-51d65f7ccff8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.274865 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-inventory" (OuterVolumeSpecName: "inventory") pod "5488228f-8fc0-46cb-be89-51d65f7ccff8" (UID: "5488228f-8fc0-46cb-be89-51d65f7ccff8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.337236 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.337311 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.337341 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzn8l\" (UniqueName: \"kubernetes.io/projected/5488228f-8fc0-46cb-be89-51d65f7ccff8-kube-api-access-jzn8l\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.337373 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5488228f-8fc0-46cb-be89-51d65f7ccff8-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.588616 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" event={"ID":"5488228f-8fc0-46cb-be89-51d65f7ccff8","Type":"ContainerDied","Data":"c88da6d7deb3f7d418f7af615e7752c5fca795e48f54f34c856119cddb0ce50e"} Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.589177 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c88da6d7deb3f7d418f7af615e7752c5fca795e48f54f34c856119cddb0ce50e" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.588711 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.771962 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f"] Oct 02 07:52:57 crc kubenswrapper[4960]: E1002 07:52:57.772440 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5488228f-8fc0-46cb-be89-51d65f7ccff8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.772460 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5488228f-8fc0-46cb-be89-51d65f7ccff8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.772648 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5488228f-8fc0-46cb-be89-51d65f7ccff8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.773420 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.780245 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.780294 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.781046 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.781168 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.792527 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.800186 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f"] Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.851991 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.852257 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.852348 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.852672 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x8m5\" (UniqueName: \"kubernetes.io/projected/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-kube-api-access-5x8m5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.954663 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x8m5\" (UniqueName: \"kubernetes.io/projected/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-kube-api-access-5x8m5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.954785 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.954827 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.954853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.959462 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.962591 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.970407 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:57 crc kubenswrapper[4960]: I1002 07:52:57.977532 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x8m5\" (UniqueName: \"kubernetes.io/projected/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-kube-api-access-5x8m5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6xk6f\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:58 crc kubenswrapper[4960]: I1002 07:52:58.094537 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:52:58 crc kubenswrapper[4960]: I1002 07:52:58.700294 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f"] Oct 02 07:52:58 crc kubenswrapper[4960]: W1002 07:52:58.720082 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7f8ad39_e87c_435f_b72b_3c59ac00f2b1.slice/crio-655b2f090f7d884ad498520823537d51aef63fa5fda0d504c3c976fea73d466d WatchSource:0}: Error finding container 655b2f090f7d884ad498520823537d51aef63fa5fda0d504c3c976fea73d466d: Status 404 returned error can't find the container with id 655b2f090f7d884ad498520823537d51aef63fa5fda0d504c3c976fea73d466d Oct 02 07:52:59 crc kubenswrapper[4960]: I1002 07:52:59.150395 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:52:59 crc kubenswrapper[4960]: I1002 07:52:59.150932 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:52:59 crc kubenswrapper[4960]: I1002 07:52:59.607453 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" event={"ID":"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1","Type":"ContainerStarted","Data":"261b53d34530d1556b1173e63d699e03e7c0f213165ac0b895c22a11595fb039"} Oct 02 07:52:59 crc kubenswrapper[4960]: I1002 07:52:59.608087 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" event={"ID":"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1","Type":"ContainerStarted","Data":"655b2f090f7d884ad498520823537d51aef63fa5fda0d504c3c976fea73d466d"} Oct 02 07:52:59 crc kubenswrapper[4960]: I1002 07:52:59.631147 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" podStartSLOduration=2.117901033 podStartE2EDuration="2.631123316s" podCreationTimestamp="2025-10-02 07:52:57 +0000 UTC" firstStartedPulling="2025-10-02 07:52:58.725552518 +0000 UTC m=+2199.757498815" lastFinishedPulling="2025-10-02 07:52:59.238774811 +0000 UTC m=+2200.270721098" observedRunningTime="2025-10-02 07:52:59.628417405 +0000 UTC m=+2200.660363692" watchObservedRunningTime="2025-10-02 07:52:59.631123316 +0000 UTC m=+2200.663069603" Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.150101 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.151121 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.151191 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.152200 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.152285 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" gracePeriod=600 Oct 02 07:53:29 crc kubenswrapper[4960]: E1002 07:53:29.280885 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.938562 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" exitCode=0 Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.938627 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755"} Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.938680 4960 scope.go:117] "RemoveContainer" containerID="8719c408af0adf4927f760c6fbcd579ca45bc4eb06efea39eb9e845e3a7c5024" Oct 02 07:53:29 crc kubenswrapper[4960]: I1002 07:53:29.940043 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:53:29 crc kubenswrapper[4960]: E1002 07:53:29.941427 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:53:37 crc kubenswrapper[4960]: I1002 07:53:37.031041 4960 generic.go:334] "Generic (PLEG): container finished" podID="e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" containerID="261b53d34530d1556b1173e63d699e03e7c0f213165ac0b895c22a11595fb039" exitCode=0 Oct 02 07:53:37 crc kubenswrapper[4960]: I1002 07:53:37.031167 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" event={"ID":"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1","Type":"ContainerDied","Data":"261b53d34530d1556b1173e63d699e03e7c0f213165ac0b895c22a11595fb039"} Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.583646 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.761926 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x8m5\" (UniqueName: \"kubernetes.io/projected/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-kube-api-access-5x8m5\") pod \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.762061 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-inventory\") pod \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.762137 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ssh-key\") pod \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.762198 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ceph\") pod \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\" (UID: \"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1\") " Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.768686 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ceph" (OuterVolumeSpecName: "ceph") pod "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" (UID: "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.775346 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-kube-api-access-5x8m5" (OuterVolumeSpecName: "kube-api-access-5x8m5") pod "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" (UID: "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1"). InnerVolumeSpecName "kube-api-access-5x8m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.789858 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" (UID: "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.820193 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-inventory" (OuterVolumeSpecName: "inventory") pod "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" (UID: "e7f8ad39-e87c-435f-b72b-3c59ac00f2b1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.865295 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.865335 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.865350 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x8m5\" (UniqueName: \"kubernetes.io/projected/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-kube-api-access-5x8m5\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:38 crc kubenswrapper[4960]: I1002 07:53:38.865366 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7f8ad39-e87c-435f-b72b-3c59ac00f2b1-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.052303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" event={"ID":"e7f8ad39-e87c-435f-b72b-3c59ac00f2b1","Type":"ContainerDied","Data":"655b2f090f7d884ad498520823537d51aef63fa5fda0d504c3c976fea73d466d"} Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.052343 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="655b2f090f7d884ad498520823537d51aef63fa5fda0d504c3c976fea73d466d" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.052373 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6xk6f" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.173504 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z"] Oct 02 07:53:39 crc kubenswrapper[4960]: E1002 07:53:39.174062 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.174086 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.174366 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f8ad39-e87c-435f-b72b-3c59ac00f2b1" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.175216 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.184388 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.186093 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.186477 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.186926 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.187014 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.188741 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z"] Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.273281 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxzh\" (UniqueName: \"kubernetes.io/projected/05979476-afd0-404e-87ed-23b6706381f2-kube-api-access-cpxzh\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.273465 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.273511 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.273671 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.375698 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxzh\" (UniqueName: \"kubernetes.io/projected/05979476-afd0-404e-87ed-23b6706381f2-kube-api-access-cpxzh\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.375831 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.375869 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.375959 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.380196 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.380490 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.389741 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.391967 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxzh\" (UniqueName: \"kubernetes.io/projected/05979476-afd0-404e-87ed-23b6706381f2-kube-api-access-cpxzh\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:39 crc kubenswrapper[4960]: I1002 07:53:39.500303 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:40 crc kubenswrapper[4960]: I1002 07:53:40.046251 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z"] Oct 02 07:53:40 crc kubenswrapper[4960]: I1002 07:53:40.061057 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" event={"ID":"05979476-afd0-404e-87ed-23b6706381f2","Type":"ContainerStarted","Data":"35274abed815ec14d7d3a78faa131afae3530a0be2272ed562cf319650a41eba"} Oct 02 07:53:41 crc kubenswrapper[4960]: I1002 07:53:41.075041 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" event={"ID":"05979476-afd0-404e-87ed-23b6706381f2","Type":"ContainerStarted","Data":"d913d9bf792fca87e7db0a486dd30a7ae2204a69c48bea997728b439080eaf92"} Oct 02 07:53:41 crc kubenswrapper[4960]: I1002 07:53:41.096451 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" podStartSLOduration=1.698657152 podStartE2EDuration="2.096431029s" podCreationTimestamp="2025-10-02 07:53:39 +0000 UTC" firstStartedPulling="2025-10-02 07:53:40.050109542 +0000 UTC m=+2241.082055839" lastFinishedPulling="2025-10-02 07:53:40.447883419 +0000 UTC m=+2241.479829716" observedRunningTime="2025-10-02 07:53:41.090834412 +0000 UTC m=+2242.122780699" watchObservedRunningTime="2025-10-02 07:53:41.096431029 +0000 UTC m=+2242.128377316" Oct 02 07:53:45 crc kubenswrapper[4960]: I1002 07:53:45.110777 4960 generic.go:334] "Generic (PLEG): container finished" podID="05979476-afd0-404e-87ed-23b6706381f2" containerID="d913d9bf792fca87e7db0a486dd30a7ae2204a69c48bea997728b439080eaf92" exitCode=0 Oct 02 07:53:45 crc kubenswrapper[4960]: I1002 07:53:45.110848 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" event={"ID":"05979476-afd0-404e-87ed-23b6706381f2","Type":"ContainerDied","Data":"d913d9bf792fca87e7db0a486dd30a7ae2204a69c48bea997728b439080eaf92"} Oct 02 07:53:45 crc kubenswrapper[4960]: I1002 07:53:45.330711 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:53:45 crc kubenswrapper[4960]: E1002 07:53:45.331008 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.485089 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.658422 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpxzh\" (UniqueName: \"kubernetes.io/projected/05979476-afd0-404e-87ed-23b6706381f2-kube-api-access-cpxzh\") pod \"05979476-afd0-404e-87ed-23b6706381f2\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.658515 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ceph\") pod \"05979476-afd0-404e-87ed-23b6706381f2\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.658635 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-inventory\") pod \"05979476-afd0-404e-87ed-23b6706381f2\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.659747 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ssh-key\") pod \"05979476-afd0-404e-87ed-23b6706381f2\" (UID: \"05979476-afd0-404e-87ed-23b6706381f2\") " Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.666197 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05979476-afd0-404e-87ed-23b6706381f2-kube-api-access-cpxzh" (OuterVolumeSpecName: "kube-api-access-cpxzh") pod "05979476-afd0-404e-87ed-23b6706381f2" (UID: "05979476-afd0-404e-87ed-23b6706381f2"). InnerVolumeSpecName "kube-api-access-cpxzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.667828 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ceph" (OuterVolumeSpecName: "ceph") pod "05979476-afd0-404e-87ed-23b6706381f2" (UID: "05979476-afd0-404e-87ed-23b6706381f2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.695110 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-inventory" (OuterVolumeSpecName: "inventory") pod "05979476-afd0-404e-87ed-23b6706381f2" (UID: "05979476-afd0-404e-87ed-23b6706381f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.720513 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "05979476-afd0-404e-87ed-23b6706381f2" (UID: "05979476-afd0-404e-87ed-23b6706381f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.763098 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpxzh\" (UniqueName: \"kubernetes.io/projected/05979476-afd0-404e-87ed-23b6706381f2-kube-api-access-cpxzh\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.763144 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.763160 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:46 crc kubenswrapper[4960]: I1002 07:53:46.763172 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05979476-afd0-404e-87ed-23b6706381f2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.130986 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" event={"ID":"05979476-afd0-404e-87ed-23b6706381f2","Type":"ContainerDied","Data":"35274abed815ec14d7d3a78faa131afae3530a0be2272ed562cf319650a41eba"} Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.131065 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35274abed815ec14d7d3a78faa131afae3530a0be2272ed562cf319650a41eba" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.131062 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.219154 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw"] Oct 02 07:53:47 crc kubenswrapper[4960]: E1002 07:53:47.219620 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05979476-afd0-404e-87ed-23b6706381f2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.219642 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="05979476-afd0-404e-87ed-23b6706381f2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.219964 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="05979476-afd0-404e-87ed-23b6706381f2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.220817 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.223649 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.223757 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.224057 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.227957 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.232270 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.242655 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw"] Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.375605 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.375688 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc7vd\" (UniqueName: \"kubernetes.io/projected/7fe7ff43-8e73-408d-bd2d-262e48941ccd-kube-api-access-nc7vd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.376509 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.376743 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.479206 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.479325 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc7vd\" (UniqueName: \"kubernetes.io/projected/7fe7ff43-8e73-408d-bd2d-262e48941ccd-kube-api-access-nc7vd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.479399 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.479476 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.484559 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.484664 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.486334 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.509843 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc7vd\" (UniqueName: \"kubernetes.io/projected/7fe7ff43-8e73-408d-bd2d-262e48941ccd-kube-api-access-nc7vd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:47 crc kubenswrapper[4960]: I1002 07:53:47.541011 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:53:48 crc kubenswrapper[4960]: I1002 07:53:48.150430 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw"] Oct 02 07:53:49 crc kubenswrapper[4960]: I1002 07:53:49.154050 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" event={"ID":"7fe7ff43-8e73-408d-bd2d-262e48941ccd","Type":"ContainerStarted","Data":"be446a118e0fd065c2363355833427bde7d59646235a0574af4fbc7a3aac29bc"} Oct 02 07:53:49 crc kubenswrapper[4960]: I1002 07:53:49.154534 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" event={"ID":"7fe7ff43-8e73-408d-bd2d-262e48941ccd","Type":"ContainerStarted","Data":"2ab96c50db8911b00b8c7e77458ddc0df7c6fe89c902aa725333182629bbb16e"} Oct 02 07:53:49 crc kubenswrapper[4960]: I1002 07:53:49.186270 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" podStartSLOduration=1.797793837 podStartE2EDuration="2.18623827s" podCreationTimestamp="2025-10-02 07:53:47 +0000 UTC" firstStartedPulling="2025-10-02 07:53:48.163147942 +0000 UTC m=+2249.195094249" lastFinishedPulling="2025-10-02 07:53:48.551592365 +0000 UTC m=+2249.583538682" observedRunningTime="2025-10-02 07:53:49.176538514 +0000 UTC m=+2250.208484841" watchObservedRunningTime="2025-10-02 07:53:49.18623827 +0000 UTC m=+2250.218184567" Oct 02 07:54:00 crc kubenswrapper[4960]: I1002 07:54:00.339460 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:54:00 crc kubenswrapper[4960]: E1002 07:54:00.341263 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:54:14 crc kubenswrapper[4960]: I1002 07:54:14.330913 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:54:14 crc kubenswrapper[4960]: E1002 07:54:14.332377 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:54:26 crc kubenswrapper[4960]: I1002 07:54:26.331014 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:54:26 crc kubenswrapper[4960]: E1002 07:54:26.333314 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:54:35 crc kubenswrapper[4960]: I1002 07:54:35.672461 4960 generic.go:334] "Generic (PLEG): container finished" podID="7fe7ff43-8e73-408d-bd2d-262e48941ccd" containerID="be446a118e0fd065c2363355833427bde7d59646235a0574af4fbc7a3aac29bc" exitCode=0 Oct 02 07:54:35 crc kubenswrapper[4960]: I1002 07:54:35.672558 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" event={"ID":"7fe7ff43-8e73-408d-bd2d-262e48941ccd","Type":"ContainerDied","Data":"be446a118e0fd065c2363355833427bde7d59646235a0574af4fbc7a3aac29bc"} Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.167616 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.280825 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-inventory\") pod \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.281126 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ceph\") pod \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.281275 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ssh-key\") pod \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.281320 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc7vd\" (UniqueName: \"kubernetes.io/projected/7fe7ff43-8e73-408d-bd2d-262e48941ccd-kube-api-access-nc7vd\") pod \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\" (UID: \"7fe7ff43-8e73-408d-bd2d-262e48941ccd\") " Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.288797 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe7ff43-8e73-408d-bd2d-262e48941ccd-kube-api-access-nc7vd" (OuterVolumeSpecName: "kube-api-access-nc7vd") pod "7fe7ff43-8e73-408d-bd2d-262e48941ccd" (UID: "7fe7ff43-8e73-408d-bd2d-262e48941ccd"). InnerVolumeSpecName "kube-api-access-nc7vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.289228 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ceph" (OuterVolumeSpecName: "ceph") pod "7fe7ff43-8e73-408d-bd2d-262e48941ccd" (UID: "7fe7ff43-8e73-408d-bd2d-262e48941ccd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.311540 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-inventory" (OuterVolumeSpecName: "inventory") pod "7fe7ff43-8e73-408d-bd2d-262e48941ccd" (UID: "7fe7ff43-8e73-408d-bd2d-262e48941ccd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.326890 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fe7ff43-8e73-408d-bd2d-262e48941ccd" (UID: "7fe7ff43-8e73-408d-bd2d-262e48941ccd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.386771 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.386821 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc7vd\" (UniqueName: \"kubernetes.io/projected/7fe7ff43-8e73-408d-bd2d-262e48941ccd-kube-api-access-nc7vd\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.386834 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.386851 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fe7ff43-8e73-408d-bd2d-262e48941ccd-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.692787 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" event={"ID":"7fe7ff43-8e73-408d-bd2d-262e48941ccd","Type":"ContainerDied","Data":"2ab96c50db8911b00b8c7e77458ddc0df7c6fe89c902aa725333182629bbb16e"} Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.692831 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ab96c50db8911b00b8c7e77458ddc0df7c6fe89c902aa725333182629bbb16e" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.692936 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.817894 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-s7hvl"] Oct 02 07:54:37 crc kubenswrapper[4960]: E1002 07:54:37.818632 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe7ff43-8e73-408d-bd2d-262e48941ccd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.818667 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe7ff43-8e73-408d-bd2d-262e48941ccd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.818936 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe7ff43-8e73-408d-bd2d-262e48941ccd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.819863 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.823239 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.823574 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.825093 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.825388 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.825633 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.834839 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-s7hvl"] Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.900572 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.900950 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ceph\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.901228 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:37 crc kubenswrapper[4960]: I1002 07:54:37.901284 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6xwn\" (UniqueName: \"kubernetes.io/projected/843c7228-62d1-4839-9459-3de9d174f70a-kube-api-access-j6xwn\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.003862 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ceph\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.004018 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.004053 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6xwn\" (UniqueName: \"kubernetes.io/projected/843c7228-62d1-4839-9459-3de9d174f70a-kube-api-access-j6xwn\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.004131 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.010450 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.010531 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ceph\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.016693 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.023160 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6xwn\" (UniqueName: \"kubernetes.io/projected/843c7228-62d1-4839-9459-3de9d174f70a-kube-api-access-j6xwn\") pod \"ssh-known-hosts-edpm-deployment-s7hvl\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.147462 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:38 crc kubenswrapper[4960]: I1002 07:54:38.725691 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-s7hvl"] Oct 02 07:54:39 crc kubenswrapper[4960]: I1002 07:54:39.713649 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" event={"ID":"843c7228-62d1-4839-9459-3de9d174f70a","Type":"ContainerStarted","Data":"105df5c99dd074ebacc542a72129dfa26ae8a82385e04bae532f05b9ab9b1f9d"} Oct 02 07:54:39 crc kubenswrapper[4960]: I1002 07:54:39.714373 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" event={"ID":"843c7228-62d1-4839-9459-3de9d174f70a","Type":"ContainerStarted","Data":"bb0f0e4031af6f348913a69a3fd8cf83c2d5b928aaeaad2fda0dc52f80a379a8"} Oct 02 07:54:39 crc kubenswrapper[4960]: I1002 07:54:39.742153 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" podStartSLOduration=2.251671005 podStartE2EDuration="2.742123441s" podCreationTimestamp="2025-10-02 07:54:37 +0000 UTC" firstStartedPulling="2025-10-02 07:54:38.730369772 +0000 UTC m=+2299.762316059" lastFinishedPulling="2025-10-02 07:54:39.220822168 +0000 UTC m=+2300.252768495" observedRunningTime="2025-10-02 07:54:39.735123306 +0000 UTC m=+2300.767069593" watchObservedRunningTime="2025-10-02 07:54:39.742123441 +0000 UTC m=+2300.774069738" Oct 02 07:54:41 crc kubenswrapper[4960]: I1002 07:54:41.330113 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:54:41 crc kubenswrapper[4960]: E1002 07:54:41.331067 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:54:50 crc kubenswrapper[4960]: I1002 07:54:50.820300 4960 generic.go:334] "Generic (PLEG): container finished" podID="843c7228-62d1-4839-9459-3de9d174f70a" containerID="105df5c99dd074ebacc542a72129dfa26ae8a82385e04bae532f05b9ab9b1f9d" exitCode=0 Oct 02 07:54:50 crc kubenswrapper[4960]: I1002 07:54:50.821121 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" event={"ID":"843c7228-62d1-4839-9459-3de9d174f70a","Type":"ContainerDied","Data":"105df5c99dd074ebacc542a72129dfa26ae8a82385e04bae532f05b9ab9b1f9d"} Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.232110 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.305138 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ceph\") pod \"843c7228-62d1-4839-9459-3de9d174f70a\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.305265 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ssh-key-openstack-edpm-ipam\") pod \"843c7228-62d1-4839-9459-3de9d174f70a\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.305300 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-inventory-0\") pod \"843c7228-62d1-4839-9459-3de9d174f70a\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.305427 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6xwn\" (UniqueName: \"kubernetes.io/projected/843c7228-62d1-4839-9459-3de9d174f70a-kube-api-access-j6xwn\") pod \"843c7228-62d1-4839-9459-3de9d174f70a\" (UID: \"843c7228-62d1-4839-9459-3de9d174f70a\") " Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.311658 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/843c7228-62d1-4839-9459-3de9d174f70a-kube-api-access-j6xwn" (OuterVolumeSpecName: "kube-api-access-j6xwn") pod "843c7228-62d1-4839-9459-3de9d174f70a" (UID: "843c7228-62d1-4839-9459-3de9d174f70a"). InnerVolumeSpecName "kube-api-access-j6xwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.313395 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ceph" (OuterVolumeSpecName: "ceph") pod "843c7228-62d1-4839-9459-3de9d174f70a" (UID: "843c7228-62d1-4839-9459-3de9d174f70a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.336343 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "843c7228-62d1-4839-9459-3de9d174f70a" (UID: "843c7228-62d1-4839-9459-3de9d174f70a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.357146 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "843c7228-62d1-4839-9459-3de9d174f70a" (UID: "843c7228-62d1-4839-9459-3de9d174f70a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.408301 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.408346 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.408361 4960 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/843c7228-62d1-4839-9459-3de9d174f70a-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.408375 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6xwn\" (UniqueName: \"kubernetes.io/projected/843c7228-62d1-4839-9459-3de9d174f70a-kube-api-access-j6xwn\") on node \"crc\" DevicePath \"\"" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.849561 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" event={"ID":"843c7228-62d1-4839-9459-3de9d174f70a","Type":"ContainerDied","Data":"bb0f0e4031af6f348913a69a3fd8cf83c2d5b928aaeaad2fda0dc52f80a379a8"} Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.849922 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb0f0e4031af6f348913a69a3fd8cf83c2d5b928aaeaad2fda0dc52f80a379a8" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.849704 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-s7hvl" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.929348 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr"] Oct 02 07:54:52 crc kubenswrapper[4960]: E1002 07:54:52.929898 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="843c7228-62d1-4839-9459-3de9d174f70a" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.929922 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="843c7228-62d1-4839-9459-3de9d174f70a" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.930203 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="843c7228-62d1-4839-9459-3de9d174f70a" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.931068 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.937137 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.937141 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.937267 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.937332 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.937498 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:54:52 crc kubenswrapper[4960]: I1002 07:54:52.942746 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr"] Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.020094 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.020149 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmp56\" (UniqueName: \"kubernetes.io/projected/5066f055-a35b-4ef7-a935-038d15d9838c-kube-api-access-jmp56\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.020251 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.020300 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.122063 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.122121 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.122234 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.122264 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmp56\" (UniqueName: \"kubernetes.io/projected/5066f055-a35b-4ef7-a935-038d15d9838c-kube-api-access-jmp56\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.127297 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.127473 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.127691 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.143394 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmp56\" (UniqueName: \"kubernetes.io/projected/5066f055-a35b-4ef7-a935-038d15d9838c-kube-api-access-jmp56\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-phrfr\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.258499 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.838580 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr"] Oct 02 07:54:53 crc kubenswrapper[4960]: I1002 07:54:53.866272 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" event={"ID":"5066f055-a35b-4ef7-a935-038d15d9838c","Type":"ContainerStarted","Data":"12cbe25663a1c9a086892a63af15e9cf98913dc98c2aa64d6716f04f1a76063d"} Oct 02 07:54:54 crc kubenswrapper[4960]: I1002 07:54:54.876329 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" event={"ID":"5066f055-a35b-4ef7-a935-038d15d9838c","Type":"ContainerStarted","Data":"38bcefc57005381bf74a708a0837849bc477ed408e8d61b093906649cfac95e1"} Oct 02 07:54:54 crc kubenswrapper[4960]: I1002 07:54:54.914214 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" podStartSLOduration=2.332433986 podStartE2EDuration="2.914181574s" podCreationTimestamp="2025-10-02 07:54:52 +0000 UTC" firstStartedPulling="2025-10-02 07:54:53.852799665 +0000 UTC m=+2314.884745962" lastFinishedPulling="2025-10-02 07:54:54.434547273 +0000 UTC m=+2315.466493550" observedRunningTime="2025-10-02 07:54:54.903723088 +0000 UTC m=+2315.935669375" watchObservedRunningTime="2025-10-02 07:54:54.914181574 +0000 UTC m=+2315.946127901" Oct 02 07:54:56 crc kubenswrapper[4960]: I1002 07:54:56.331594 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:54:56 crc kubenswrapper[4960]: E1002 07:54:56.332881 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:55:02 crc kubenswrapper[4960]: I1002 07:55:02.982014 4960 generic.go:334] "Generic (PLEG): container finished" podID="5066f055-a35b-4ef7-a935-038d15d9838c" containerID="38bcefc57005381bf74a708a0837849bc477ed408e8d61b093906649cfac95e1" exitCode=0 Oct 02 07:55:02 crc kubenswrapper[4960]: I1002 07:55:02.982657 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" event={"ID":"5066f055-a35b-4ef7-a935-038d15d9838c","Type":"ContainerDied","Data":"38bcefc57005381bf74a708a0837849bc477ed408e8d61b093906649cfac95e1"} Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.422511 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.467586 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmp56\" (UniqueName: \"kubernetes.io/projected/5066f055-a35b-4ef7-a935-038d15d9838c-kube-api-access-jmp56\") pod \"5066f055-a35b-4ef7-a935-038d15d9838c\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.467638 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ceph\") pod \"5066f055-a35b-4ef7-a935-038d15d9838c\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.467716 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-inventory\") pod \"5066f055-a35b-4ef7-a935-038d15d9838c\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.467883 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ssh-key\") pod \"5066f055-a35b-4ef7-a935-038d15d9838c\" (UID: \"5066f055-a35b-4ef7-a935-038d15d9838c\") " Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.476782 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ceph" (OuterVolumeSpecName: "ceph") pod "5066f055-a35b-4ef7-a935-038d15d9838c" (UID: "5066f055-a35b-4ef7-a935-038d15d9838c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.476829 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5066f055-a35b-4ef7-a935-038d15d9838c-kube-api-access-jmp56" (OuterVolumeSpecName: "kube-api-access-jmp56") pod "5066f055-a35b-4ef7-a935-038d15d9838c" (UID: "5066f055-a35b-4ef7-a935-038d15d9838c"). InnerVolumeSpecName "kube-api-access-jmp56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.497420 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-inventory" (OuterVolumeSpecName: "inventory") pod "5066f055-a35b-4ef7-a935-038d15d9838c" (UID: "5066f055-a35b-4ef7-a935-038d15d9838c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.518509 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5066f055-a35b-4ef7-a935-038d15d9838c" (UID: "5066f055-a35b-4ef7-a935-038d15d9838c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.571045 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmp56\" (UniqueName: \"kubernetes.io/projected/5066f055-a35b-4ef7-a935-038d15d9838c-kube-api-access-jmp56\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.571082 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.571096 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:04 crc kubenswrapper[4960]: I1002 07:55:04.571111 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5066f055-a35b-4ef7-a935-038d15d9838c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.008421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" event={"ID":"5066f055-a35b-4ef7-a935-038d15d9838c","Type":"ContainerDied","Data":"12cbe25663a1c9a086892a63af15e9cf98913dc98c2aa64d6716f04f1a76063d"} Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.008586 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12cbe25663a1c9a086892a63af15e9cf98913dc98c2aa64d6716f04f1a76063d" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.008675 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-phrfr" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.119184 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4"] Oct 02 07:55:05 crc kubenswrapper[4960]: E1002 07:55:05.119776 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5066f055-a35b-4ef7-a935-038d15d9838c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.119808 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5066f055-a35b-4ef7-a935-038d15d9838c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.120208 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5066f055-a35b-4ef7-a935-038d15d9838c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.121256 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.124130 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.124830 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.124841 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.124859 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.127028 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.146664 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4"] Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.186500 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.186601 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.186662 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvqdg\" (UniqueName: \"kubernetes.io/projected/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-kube-api-access-hvqdg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.187119 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.289203 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.289374 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.289463 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.289548 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvqdg\" (UniqueName: \"kubernetes.io/projected/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-kube-api-access-hvqdg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.295049 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.295186 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.295872 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.314269 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvqdg\" (UniqueName: \"kubernetes.io/projected/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-kube-api-access-hvqdg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:05 crc kubenswrapper[4960]: I1002 07:55:05.447997 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:06 crc kubenswrapper[4960]: I1002 07:55:06.103720 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4"] Oct 02 07:55:07 crc kubenswrapper[4960]: I1002 07:55:07.035222 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" event={"ID":"defafea0-a1a3-4e6b-82ce-5b431be9ff5b","Type":"ContainerStarted","Data":"afccb56b2032ad09d41436e75f09306efa1741544c96283e2d9614be8d8f79d2"} Oct 02 07:55:07 crc kubenswrapper[4960]: I1002 07:55:07.035904 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" event={"ID":"defafea0-a1a3-4e6b-82ce-5b431be9ff5b","Type":"ContainerStarted","Data":"524b9f489663778608833f5fcee1a2eab7c611c12e0019e82f33ce03f9fec9b3"} Oct 02 07:55:07 crc kubenswrapper[4960]: I1002 07:55:07.063915 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" podStartSLOduration=1.640325485 podStartE2EDuration="2.063889974s" podCreationTimestamp="2025-10-02 07:55:05 +0000 UTC" firstStartedPulling="2025-10-02 07:55:06.126187713 +0000 UTC m=+2327.158134000" lastFinishedPulling="2025-10-02 07:55:06.549752172 +0000 UTC m=+2327.581698489" observedRunningTime="2025-10-02 07:55:07.056281393 +0000 UTC m=+2328.088227690" watchObservedRunningTime="2025-10-02 07:55:07.063889974 +0000 UTC m=+2328.095836261" Oct 02 07:55:08 crc kubenswrapper[4960]: I1002 07:55:08.330480 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:55:08 crc kubenswrapper[4960]: E1002 07:55:08.330726 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:55:17 crc kubenswrapper[4960]: I1002 07:55:17.175311 4960 generic.go:334] "Generic (PLEG): container finished" podID="defafea0-a1a3-4e6b-82ce-5b431be9ff5b" containerID="afccb56b2032ad09d41436e75f09306efa1741544c96283e2d9614be8d8f79d2" exitCode=0 Oct 02 07:55:17 crc kubenswrapper[4960]: I1002 07:55:17.175405 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" event={"ID":"defafea0-a1a3-4e6b-82ce-5b431be9ff5b","Type":"ContainerDied","Data":"afccb56b2032ad09d41436e75f09306efa1741544c96283e2d9614be8d8f79d2"} Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.698604 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.834960 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvqdg\" (UniqueName: \"kubernetes.io/projected/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-kube-api-access-hvqdg\") pod \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.835520 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ssh-key\") pod \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.835664 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ceph\") pod \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.835780 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-inventory\") pod \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\" (UID: \"defafea0-a1a3-4e6b-82ce-5b431be9ff5b\") " Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.842954 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ceph" (OuterVolumeSpecName: "ceph") pod "defafea0-a1a3-4e6b-82ce-5b431be9ff5b" (UID: "defafea0-a1a3-4e6b-82ce-5b431be9ff5b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.843127 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-kube-api-access-hvqdg" (OuterVolumeSpecName: "kube-api-access-hvqdg") pod "defafea0-a1a3-4e6b-82ce-5b431be9ff5b" (UID: "defafea0-a1a3-4e6b-82ce-5b431be9ff5b"). InnerVolumeSpecName "kube-api-access-hvqdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.864195 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "defafea0-a1a3-4e6b-82ce-5b431be9ff5b" (UID: "defafea0-a1a3-4e6b-82ce-5b431be9ff5b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.872091 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-inventory" (OuterVolumeSpecName: "inventory") pod "defafea0-a1a3-4e6b-82ce-5b431be9ff5b" (UID: "defafea0-a1a3-4e6b-82ce-5b431be9ff5b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.938580 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvqdg\" (UniqueName: \"kubernetes.io/projected/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-kube-api-access-hvqdg\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.938621 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.938636 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:18 crc kubenswrapper[4960]: I1002 07:55:18.938645 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafea0-a1a3-4e6b-82ce-5b431be9ff5b-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.200686 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" event={"ID":"defafea0-a1a3-4e6b-82ce-5b431be9ff5b","Type":"ContainerDied","Data":"524b9f489663778608833f5fcee1a2eab7c611c12e0019e82f33ce03f9fec9b3"} Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.200733 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="524b9f489663778608833f5fcee1a2eab7c611c12e0019e82f33ce03f9fec9b3" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.200830 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.334540 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q"] Oct 02 07:55:19 crc kubenswrapper[4960]: E1002 07:55:19.335133 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="defafea0-a1a3-4e6b-82ce-5b431be9ff5b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.335164 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="defafea0-a1a3-4e6b-82ce-5b431be9ff5b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.335474 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="defafea0-a1a3-4e6b-82ce-5b431be9ff5b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.336538 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.340139 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.340740 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.341541 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.341991 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.344674 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.345011 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.345237 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.345638 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.357388 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q"] Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.448348 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.448434 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449269 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449331 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449549 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449626 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449665 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449813 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449857 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449912 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.449964 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn5hj\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-kube-api-access-zn5hj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.450112 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.450165 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552552 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn5hj\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-kube-api-access-zn5hj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552633 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552665 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552751 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552786 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552830 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552854 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552903 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.552938 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.553027 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.553060 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.553093 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.553119 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.559439 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.559530 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.559558 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.560316 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.560335 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.560841 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.560859 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.561247 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.562198 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.562717 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.562862 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.563149 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.571943 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn5hj\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-kube-api-access-zn5hj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:19 crc kubenswrapper[4960]: I1002 07:55:19.671319 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:20 crc kubenswrapper[4960]: I1002 07:55:20.269550 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q"] Oct 02 07:55:20 crc kubenswrapper[4960]: I1002 07:55:20.774633 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:55:21 crc kubenswrapper[4960]: I1002 07:55:21.225900 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" event={"ID":"5aa32b54-0374-4032-b19f-d1e65557fb95","Type":"ContainerStarted","Data":"dd3a06d93ecf13107635b761159693b0a5a2f5c8ed8f2c399b7c0f01417b43f2"} Oct 02 07:55:21 crc kubenswrapper[4960]: I1002 07:55:21.225946 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" event={"ID":"5aa32b54-0374-4032-b19f-d1e65557fb95","Type":"ContainerStarted","Data":"a209ef626b0424e665a066b6d56eba6f2f3e50ac7dd6d500c838b497b76543f0"} Oct 02 07:55:21 crc kubenswrapper[4960]: I1002 07:55:21.260811 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" podStartSLOduration=1.7710571480000001 podStartE2EDuration="2.260789656s" podCreationTimestamp="2025-10-02 07:55:19 +0000 UTC" firstStartedPulling="2025-10-02 07:55:20.279939184 +0000 UTC m=+2341.311885471" lastFinishedPulling="2025-10-02 07:55:20.769671682 +0000 UTC m=+2341.801617979" observedRunningTime="2025-10-02 07:55:21.251014248 +0000 UTC m=+2342.282960555" watchObservedRunningTime="2025-10-02 07:55:21.260789656 +0000 UTC m=+2342.292735933" Oct 02 07:55:22 crc kubenswrapper[4960]: I1002 07:55:22.330332 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:55:22 crc kubenswrapper[4960]: E1002 07:55:22.331158 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:55:35 crc kubenswrapper[4960]: I1002 07:55:35.331033 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:55:35 crc kubenswrapper[4960]: E1002 07:55:35.332158 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:55:50 crc kubenswrapper[4960]: I1002 07:55:50.337220 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:55:50 crc kubenswrapper[4960]: E1002 07:55:50.338286 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:55:55 crc kubenswrapper[4960]: I1002 07:55:55.598200 4960 generic.go:334] "Generic (PLEG): container finished" podID="5aa32b54-0374-4032-b19f-d1e65557fb95" containerID="dd3a06d93ecf13107635b761159693b0a5a2f5c8ed8f2c399b7c0f01417b43f2" exitCode=0 Oct 02 07:55:55 crc kubenswrapper[4960]: I1002 07:55:55.598339 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" event={"ID":"5aa32b54-0374-4032-b19f-d1e65557fb95","Type":"ContainerDied","Data":"dd3a06d93ecf13107635b761159693b0a5a2f5c8ed8f2c399b7c0f01417b43f2"} Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.096387 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.197946 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ovn-combined-ca-bundle\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198170 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198235 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ceph\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198277 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ssh-key\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198352 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-bootstrap-combined-ca-bundle\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198391 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-neutron-metadata-combined-ca-bundle\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198484 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-repo-setup-combined-ca-bundle\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198519 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-inventory\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.198586 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-nova-combined-ca-bundle\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.205812 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.207811 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.207862 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.207867 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.207916 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ceph" (OuterVolumeSpecName: "ceph") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.208548 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.208959 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.229489 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-inventory" (OuterVolumeSpecName: "inventory") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.256606 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.300847 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.300917 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.301122 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn5hj\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-kube-api-access-zn5hj\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.301181 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-libvirt-combined-ca-bundle\") pod \"5aa32b54-0374-4032-b19f-d1e65557fb95\" (UID: \"5aa32b54-0374-4032-b19f-d1e65557fb95\") " Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302099 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302131 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302153 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302175 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302195 4960 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302214 4960 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302231 4960 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302250 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.302268 4960 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.304891 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.305474 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-kube-api-access-zn5hj" (OuterVolumeSpecName: "kube-api-access-zn5hj") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "kube-api-access-zn5hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.305548 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.307164 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5aa32b54-0374-4032-b19f-d1e65557fb95" (UID: "5aa32b54-0374-4032-b19f-d1e65557fb95"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.404433 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.404975 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.405021 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn5hj\" (UniqueName: \"kubernetes.io/projected/5aa32b54-0374-4032-b19f-d1e65557fb95-kube-api-access-zn5hj\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.405033 4960 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa32b54-0374-4032-b19f-d1e65557fb95-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.624505 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" event={"ID":"5aa32b54-0374-4032-b19f-d1e65557fb95","Type":"ContainerDied","Data":"a209ef626b0424e665a066b6d56eba6f2f3e50ac7dd6d500c838b497b76543f0"} Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.624573 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a209ef626b0424e665a066b6d56eba6f2f3e50ac7dd6d500c838b497b76543f0" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.624604 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.771267 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2"] Oct 02 07:55:57 crc kubenswrapper[4960]: E1002 07:55:57.771739 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa32b54-0374-4032-b19f-d1e65557fb95" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.771760 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa32b54-0374-4032-b19f-d1e65557fb95" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.771966 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aa32b54-0374-4032-b19f-d1e65557fb95" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.772865 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.775515 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.776182 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.776415 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.776613 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.777986 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.779699 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2"] Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.811448 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.811597 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.811673 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz5xz\" (UniqueName: \"kubernetes.io/projected/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-kube-api-access-lz5xz\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.811788 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.913138 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.913212 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz5xz\" (UniqueName: \"kubernetes.io/projected/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-kube-api-access-lz5xz\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.913263 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.913326 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.920323 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.920813 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.920872 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:57 crc kubenswrapper[4960]: I1002 07:55:57.933932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz5xz\" (UniqueName: \"kubernetes.io/projected/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-kube-api-access-lz5xz\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:58 crc kubenswrapper[4960]: I1002 07:55:58.145744 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:55:58 crc kubenswrapper[4960]: I1002 07:55:58.736671 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2"] Oct 02 07:55:58 crc kubenswrapper[4960]: I1002 07:55:58.753249 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:55:59 crc kubenswrapper[4960]: I1002 07:55:59.652070 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" event={"ID":"363e150c-f1f7-4cd2-9916-e8724ceb0f9a","Type":"ContainerStarted","Data":"15963421f5821db6a61e87710d45f175c4bc6ba0b9c0c13849a3521456770b20"} Oct 02 07:55:59 crc kubenswrapper[4960]: I1002 07:55:59.652451 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" event={"ID":"363e150c-f1f7-4cd2-9916-e8724ceb0f9a","Type":"ContainerStarted","Data":"60d47f78715b03d9962f6ec26e86c77cbc5826918a3119350092783eabdf4628"} Oct 02 07:55:59 crc kubenswrapper[4960]: I1002 07:55:59.674964 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" podStartSLOduration=2.112628972 podStartE2EDuration="2.674940647s" podCreationTimestamp="2025-10-02 07:55:57 +0000 UTC" firstStartedPulling="2025-10-02 07:55:58.753037902 +0000 UTC m=+2379.784984189" lastFinishedPulling="2025-10-02 07:55:59.315349537 +0000 UTC m=+2380.347295864" observedRunningTime="2025-10-02 07:55:59.672525153 +0000 UTC m=+2380.704471460" watchObservedRunningTime="2025-10-02 07:55:59.674940647 +0000 UTC m=+2380.706886944" Oct 02 07:56:04 crc kubenswrapper[4960]: I1002 07:56:04.330462 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:56:04 crc kubenswrapper[4960]: E1002 07:56:04.331386 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:56:05 crc kubenswrapper[4960]: I1002 07:56:05.721246 4960 generic.go:334] "Generic (PLEG): container finished" podID="363e150c-f1f7-4cd2-9916-e8724ceb0f9a" containerID="15963421f5821db6a61e87710d45f175c4bc6ba0b9c0c13849a3521456770b20" exitCode=0 Oct 02 07:56:05 crc kubenswrapper[4960]: I1002 07:56:05.721364 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" event={"ID":"363e150c-f1f7-4cd2-9916-e8724ceb0f9a","Type":"ContainerDied","Data":"15963421f5821db6a61e87710d45f175c4bc6ba0b9c0c13849a3521456770b20"} Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.244463 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.433264 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-inventory\") pod \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.433444 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ssh-key\") pod \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.433533 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz5xz\" (UniqueName: \"kubernetes.io/projected/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-kube-api-access-lz5xz\") pod \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.433649 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ceph\") pod \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\" (UID: \"363e150c-f1f7-4cd2-9916-e8724ceb0f9a\") " Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.441706 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ceph" (OuterVolumeSpecName: "ceph") pod "363e150c-f1f7-4cd2-9916-e8724ceb0f9a" (UID: "363e150c-f1f7-4cd2-9916-e8724ceb0f9a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.442505 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-kube-api-access-lz5xz" (OuterVolumeSpecName: "kube-api-access-lz5xz") pod "363e150c-f1f7-4cd2-9916-e8724ceb0f9a" (UID: "363e150c-f1f7-4cd2-9916-e8724ceb0f9a"). InnerVolumeSpecName "kube-api-access-lz5xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.467089 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-inventory" (OuterVolumeSpecName: "inventory") pod "363e150c-f1f7-4cd2-9916-e8724ceb0f9a" (UID: "363e150c-f1f7-4cd2-9916-e8724ceb0f9a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.468111 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "363e150c-f1f7-4cd2-9916-e8724ceb0f9a" (UID: "363e150c-f1f7-4cd2-9916-e8724ceb0f9a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.537383 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.537453 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.537476 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz5xz\" (UniqueName: \"kubernetes.io/projected/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-kube-api-access-lz5xz\") on node \"crc\" DevicePath \"\"" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.537500 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/363e150c-f1f7-4cd2-9916-e8724ceb0f9a-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.749343 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" event={"ID":"363e150c-f1f7-4cd2-9916-e8724ceb0f9a","Type":"ContainerDied","Data":"60d47f78715b03d9962f6ec26e86c77cbc5826918a3119350092783eabdf4628"} Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.749788 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60d47f78715b03d9962f6ec26e86c77cbc5826918a3119350092783eabdf4628" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.749910 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.977035 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm"] Oct 02 07:56:07 crc kubenswrapper[4960]: E1002 07:56:07.977628 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363e150c-f1f7-4cd2-9916-e8724ceb0f9a" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.977651 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="363e150c-f1f7-4cd2-9916-e8724ceb0f9a" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.977881 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="363e150c-f1f7-4cd2-9916-e8724ceb0f9a" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.978966 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.983670 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.983945 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.984125 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.984281 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.984415 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.984539 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:56:07 crc kubenswrapper[4960]: I1002 07:56:07.988343 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm"] Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.151051 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.151110 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.151138 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.151646 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.152715 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.153117 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqz2w\" (UniqueName: \"kubernetes.io/projected/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-kube-api-access-vqz2w\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.256023 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.256155 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.256267 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqz2w\" (UniqueName: \"kubernetes.io/projected/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-kube-api-access-vqz2w\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.256385 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.256439 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.256509 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.257279 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.263663 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.263738 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.263842 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.266029 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.290735 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqz2w\" (UniqueName: \"kubernetes.io/projected/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-kube-api-access-vqz2w\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gdjnm\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:08 crc kubenswrapper[4960]: I1002 07:56:08.335272 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:56:09 crc kubenswrapper[4960]: I1002 07:56:09.007548 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm"] Oct 02 07:56:09 crc kubenswrapper[4960]: I1002 07:56:09.773420 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" event={"ID":"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3","Type":"ContainerStarted","Data":"caf6e9cb0da208c907665be8ff95ac34cf9460e7012098f708e5641f0f94f5ef"} Oct 02 07:56:10 crc kubenswrapper[4960]: I1002 07:56:10.815397 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" event={"ID":"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3","Type":"ContainerStarted","Data":"81faebf265aa33f125d8c4424f638b1e4632d6361016358bd8305ce2d8630cae"} Oct 02 07:56:10 crc kubenswrapper[4960]: I1002 07:56:10.856032 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" podStartSLOduration=3.264774217 podStartE2EDuration="3.855952085s" podCreationTimestamp="2025-10-02 07:56:07 +0000 UTC" firstStartedPulling="2025-10-02 07:56:09.006431135 +0000 UTC m=+2390.038377422" lastFinishedPulling="2025-10-02 07:56:09.597608973 +0000 UTC m=+2390.629555290" observedRunningTime="2025-10-02 07:56:10.843653261 +0000 UTC m=+2391.875599558" watchObservedRunningTime="2025-10-02 07:56:10.855952085 +0000 UTC m=+2391.887898382" Oct 02 07:56:15 crc kubenswrapper[4960]: I1002 07:56:15.331246 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:56:15 crc kubenswrapper[4960]: E1002 07:56:15.334306 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:56:29 crc kubenswrapper[4960]: I1002 07:56:29.329908 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:56:29 crc kubenswrapper[4960]: E1002 07:56:29.331051 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:56:41 crc kubenswrapper[4960]: I1002 07:56:41.330607 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:56:41 crc kubenswrapper[4960]: E1002 07:56:41.331588 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:56:52 crc kubenswrapper[4960]: I1002 07:56:52.330951 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:56:52 crc kubenswrapper[4960]: E1002 07:56:52.332032 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:57:03 crc kubenswrapper[4960]: I1002 07:57:03.330475 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:57:03 crc kubenswrapper[4960]: E1002 07:57:03.331807 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:57:16 crc kubenswrapper[4960]: I1002 07:57:16.331181 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:57:16 crc kubenswrapper[4960]: E1002 07:57:16.332207 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:57:26 crc kubenswrapper[4960]: I1002 07:57:26.701943 4960 generic.go:334] "Generic (PLEG): container finished" podID="ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" containerID="81faebf265aa33f125d8c4424f638b1e4632d6361016358bd8305ce2d8630cae" exitCode=0 Oct 02 07:57:26 crc kubenswrapper[4960]: I1002 07:57:26.702026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" event={"ID":"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3","Type":"ContainerDied","Data":"81faebf265aa33f125d8c4424f638b1e4632d6361016358bd8305ce2d8630cae"} Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.180344 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.279879 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-inventory\") pod \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.279922 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ceph\") pod \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.280039 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqz2w\" (UniqueName: \"kubernetes.io/projected/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-kube-api-access-vqz2w\") pod \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.280148 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovncontroller-config-0\") pod \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.280168 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ssh-key\") pod \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.280270 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovn-combined-ca-bundle\") pod \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\" (UID: \"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3\") " Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.286267 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" (UID: "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.287330 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ceph" (OuterVolumeSpecName: "ceph") pod "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" (UID: "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.287964 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-kube-api-access-vqz2w" (OuterVolumeSpecName: "kube-api-access-vqz2w") pod "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" (UID: "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3"). InnerVolumeSpecName "kube-api-access-vqz2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.310198 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-inventory" (OuterVolumeSpecName: "inventory") pod "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" (UID: "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.311634 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" (UID: "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.314562 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" (UID: "ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.383583 4960 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.383626 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.383638 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.383651 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.383667 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.383677 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqz2w\" (UniqueName: \"kubernetes.io/projected/ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3-kube-api-access-vqz2w\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.726500 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" event={"ID":"ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3","Type":"ContainerDied","Data":"caf6e9cb0da208c907665be8ff95ac34cf9460e7012098f708e5641f0f94f5ef"} Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.726901 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caf6e9cb0da208c907665be8ff95ac34cf9460e7012098f708e5641f0f94f5ef" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.726740 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gdjnm" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.868525 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw"] Oct 02 07:57:28 crc kubenswrapper[4960]: E1002 07:57:28.869061 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.869084 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.869282 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.870093 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.876905 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.877269 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.877603 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.877706 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.877798 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.877878 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.879358 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.901843 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw"] Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.995501 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.995578 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.995627 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.995662 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.995695 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvzlg\" (UniqueName: \"kubernetes.io/projected/9708c6b1-7f2f-4820-9072-12f019a1a731-kube-api-access-dvzlg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.995741 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:28 crc kubenswrapper[4960]: I1002 07:57:28.995846 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.097443 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.097568 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.097637 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.097703 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.097751 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.097792 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.097823 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvzlg\" (UniqueName: \"kubernetes.io/projected/9708c6b1-7f2f-4820-9072-12f019a1a731-kube-api-access-dvzlg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.102725 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.103620 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.103619 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.103946 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.105041 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.110598 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.123896 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvzlg\" (UniqueName: \"kubernetes.io/projected/9708c6b1-7f2f-4820-9072-12f019a1a731-kube-api-access-dvzlg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.198641 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.330129 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:57:29 crc kubenswrapper[4960]: E1002 07:57:29.330918 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:57:29 crc kubenswrapper[4960]: I1002 07:57:29.734522 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw"] Oct 02 07:57:30 crc kubenswrapper[4960]: I1002 07:57:30.755841 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" event={"ID":"9708c6b1-7f2f-4820-9072-12f019a1a731","Type":"ContainerStarted","Data":"1ef50654e75b11941bbbdde444f56cc55a7d833946c1972a98f8248ac8b21348"} Oct 02 07:57:30 crc kubenswrapper[4960]: I1002 07:57:30.756165 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" event={"ID":"9708c6b1-7f2f-4820-9072-12f019a1a731","Type":"ContainerStarted","Data":"a83349510a5a5d7192735a0e77243302c4a407882c7763492a49bb153896b605"} Oct 02 07:57:30 crc kubenswrapper[4960]: I1002 07:57:30.787268 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" podStartSLOduration=2.318237091 podStartE2EDuration="2.787212561s" podCreationTimestamp="2025-10-02 07:57:28 +0000 UTC" firstStartedPulling="2025-10-02 07:57:29.741042463 +0000 UTC m=+2470.772988750" lastFinishedPulling="2025-10-02 07:57:30.210017933 +0000 UTC m=+2471.241964220" observedRunningTime="2025-10-02 07:57:30.782421284 +0000 UTC m=+2471.814367601" watchObservedRunningTime="2025-10-02 07:57:30.787212561 +0000 UTC m=+2471.819158878" Oct 02 07:57:42 crc kubenswrapper[4960]: I1002 07:57:42.331017 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:57:42 crc kubenswrapper[4960]: E1002 07:57:42.332098 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:57:44 crc kubenswrapper[4960]: I1002 07:57:44.822323 4960 scope.go:117] "RemoveContainer" containerID="766e38df662665e37ed10a9ec73e37bf8bea511d77512a034f24a32d8bbc0f5d" Oct 02 07:57:44 crc kubenswrapper[4960]: I1002 07:57:44.856643 4960 scope.go:117] "RemoveContainer" containerID="c5112ee4291c64553463a2037fb8377c97318619afa4aad9f981071d63f0bdb9" Oct 02 07:57:44 crc kubenswrapper[4960]: I1002 07:57:44.951298 4960 scope.go:117] "RemoveContainer" containerID="066f6ef80720dba26cb30f03f8c96b3d77166e7ae3634422f568c3a973f9b72c" Oct 02 07:57:55 crc kubenswrapper[4960]: I1002 07:57:55.330890 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:57:55 crc kubenswrapper[4960]: E1002 07:57:55.332121 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:58:06 crc kubenswrapper[4960]: I1002 07:58:06.331006 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:58:06 crc kubenswrapper[4960]: E1002 07:58:06.332170 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:58:21 crc kubenswrapper[4960]: I1002 07:58:21.330989 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:58:21 crc kubenswrapper[4960]: E1002 07:58:21.332136 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 07:58:33 crc kubenswrapper[4960]: I1002 07:58:33.330907 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 07:58:34 crc kubenswrapper[4960]: I1002 07:58:34.514554 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"697e5417554915e415b18e62ccd2b91fc62b5307191461e4f59ddcb5fd447d6d"} Oct 02 07:58:34 crc kubenswrapper[4960]: I1002 07:58:34.518479 4960 generic.go:334] "Generic (PLEG): container finished" podID="9708c6b1-7f2f-4820-9072-12f019a1a731" containerID="1ef50654e75b11941bbbdde444f56cc55a7d833946c1972a98f8248ac8b21348" exitCode=0 Oct 02 07:58:34 crc kubenswrapper[4960]: I1002 07:58:34.518655 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" event={"ID":"9708c6b1-7f2f-4820-9072-12f019a1a731","Type":"ContainerDied","Data":"1ef50654e75b11941bbbdde444f56cc55a7d833946c1972a98f8248ac8b21348"} Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.106136 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.119747 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-ovn-metadata-agent-neutron-config-0\") pod \"9708c6b1-7f2f-4820-9072-12f019a1a731\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.119837 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-metadata-combined-ca-bundle\") pod \"9708c6b1-7f2f-4820-9072-12f019a1a731\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.120094 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ssh-key\") pod \"9708c6b1-7f2f-4820-9072-12f019a1a731\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.120175 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-inventory\") pod \"9708c6b1-7f2f-4820-9072-12f019a1a731\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.120209 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-nova-metadata-neutron-config-0\") pod \"9708c6b1-7f2f-4820-9072-12f019a1a731\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.120345 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ceph\") pod \"9708c6b1-7f2f-4820-9072-12f019a1a731\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.120375 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvzlg\" (UniqueName: \"kubernetes.io/projected/9708c6b1-7f2f-4820-9072-12f019a1a731-kube-api-access-dvzlg\") pod \"9708c6b1-7f2f-4820-9072-12f019a1a731\" (UID: \"9708c6b1-7f2f-4820-9072-12f019a1a731\") " Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.128532 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "9708c6b1-7f2f-4820-9072-12f019a1a731" (UID: "9708c6b1-7f2f-4820-9072-12f019a1a731"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.134811 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ceph" (OuterVolumeSpecName: "ceph") pod "9708c6b1-7f2f-4820-9072-12f019a1a731" (UID: "9708c6b1-7f2f-4820-9072-12f019a1a731"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.138837 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9708c6b1-7f2f-4820-9072-12f019a1a731-kube-api-access-dvzlg" (OuterVolumeSpecName: "kube-api-access-dvzlg") pod "9708c6b1-7f2f-4820-9072-12f019a1a731" (UID: "9708c6b1-7f2f-4820-9072-12f019a1a731"). InnerVolumeSpecName "kube-api-access-dvzlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.162811 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "9708c6b1-7f2f-4820-9072-12f019a1a731" (UID: "9708c6b1-7f2f-4820-9072-12f019a1a731"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.176543 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "9708c6b1-7f2f-4820-9072-12f019a1a731" (UID: "9708c6b1-7f2f-4820-9072-12f019a1a731"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.190120 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-inventory" (OuterVolumeSpecName: "inventory") pod "9708c6b1-7f2f-4820-9072-12f019a1a731" (UID: "9708c6b1-7f2f-4820-9072-12f019a1a731"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.196343 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9708c6b1-7f2f-4820-9072-12f019a1a731" (UID: "9708c6b1-7f2f-4820-9072-12f019a1a731"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.223992 4960 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.224028 4960 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.224040 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.224050 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.224060 4960 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.224074 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9708c6b1-7f2f-4820-9072-12f019a1a731-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.224084 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvzlg\" (UniqueName: \"kubernetes.io/projected/9708c6b1-7f2f-4820-9072-12f019a1a731-kube-api-access-dvzlg\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.562701 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" event={"ID":"9708c6b1-7f2f-4820-9072-12f019a1a731","Type":"ContainerDied","Data":"a83349510a5a5d7192735a0e77243302c4a407882c7763492a49bb153896b605"} Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.562763 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a83349510a5a5d7192735a0e77243302c4a407882c7763492a49bb153896b605" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.562818 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.701955 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc"] Oct 02 07:58:36 crc kubenswrapper[4960]: E1002 07:58:36.703256 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9708c6b1-7f2f-4820-9072-12f019a1a731" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.703280 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9708c6b1-7f2f-4820-9072-12f019a1a731" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.703489 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9708c6b1-7f2f-4820-9072-12f019a1a731" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.704654 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.711513 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.711590 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.711674 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.711666 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.711911 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.712640 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.739208 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc"] Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.740480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.740721 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg4jx\" (UniqueName: \"kubernetes.io/projected/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-kube-api-access-sg4jx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.740756 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.740809 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.740839 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.740868 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.843098 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.843231 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg4jx\" (UniqueName: \"kubernetes.io/projected/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-kube-api-access-sg4jx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.843267 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.843329 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.843370 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.843410 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.848615 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.848921 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.849371 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.849508 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.849935 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:36 crc kubenswrapper[4960]: I1002 07:58:36.862733 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg4jx\" (UniqueName: \"kubernetes.io/projected/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-kube-api-access-sg4jx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:37 crc kubenswrapper[4960]: I1002 07:58:37.030590 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 07:58:37 crc kubenswrapper[4960]: I1002 07:58:37.442778 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc"] Oct 02 07:58:37 crc kubenswrapper[4960]: W1002 07:58:37.450561 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2f4ac6d_fe72_4957_bcdb_983b63fa172b.slice/crio-ff060e1d68c875fadb684cf148c478866afb54332f2a7db1c6252d88157fa998 WatchSource:0}: Error finding container ff060e1d68c875fadb684cf148c478866afb54332f2a7db1c6252d88157fa998: Status 404 returned error can't find the container with id ff060e1d68c875fadb684cf148c478866afb54332f2a7db1c6252d88157fa998 Oct 02 07:58:37 crc kubenswrapper[4960]: I1002 07:58:37.574894 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" event={"ID":"f2f4ac6d-fe72-4957-bcdb-983b63fa172b","Type":"ContainerStarted","Data":"ff060e1d68c875fadb684cf148c478866afb54332f2a7db1c6252d88157fa998"} Oct 02 07:58:38 crc kubenswrapper[4960]: I1002 07:58:38.590812 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" event={"ID":"f2f4ac6d-fe72-4957-bcdb-983b63fa172b","Type":"ContainerStarted","Data":"db5f455cf4a4449f523f2053ba51636bced5c48cc1b015aedd7d3cc50a6aa0ab"} Oct 02 07:58:38 crc kubenswrapper[4960]: I1002 07:58:38.630088 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" podStartSLOduration=2.050367312 podStartE2EDuration="2.630066572s" podCreationTimestamp="2025-10-02 07:58:36 +0000 UTC" firstStartedPulling="2025-10-02 07:58:37.453960825 +0000 UTC m=+2538.485907112" lastFinishedPulling="2025-10-02 07:58:38.033660085 +0000 UTC m=+2539.065606372" observedRunningTime="2025-10-02 07:58:38.623060074 +0000 UTC m=+2539.655006361" watchObservedRunningTime="2025-10-02 07:58:38.630066572 +0000 UTC m=+2539.662012859" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.149996 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz"] Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.152226 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.154491 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.155485 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.159262 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz"] Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.179744 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5d666c8-3d64-4297-89eb-07c7e0a316e6-secret-volume\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.180168 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b5h4\" (UniqueName: \"kubernetes.io/projected/f5d666c8-3d64-4297-89eb-07c7e0a316e6-kube-api-access-4b5h4\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.180356 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5d666c8-3d64-4297-89eb-07c7e0a316e6-config-volume\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.281403 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5d666c8-3d64-4297-89eb-07c7e0a316e6-config-volume\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.281737 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5d666c8-3d64-4297-89eb-07c7e0a316e6-secret-volume\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.281833 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b5h4\" (UniqueName: \"kubernetes.io/projected/f5d666c8-3d64-4297-89eb-07c7e0a316e6-kube-api-access-4b5h4\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.282493 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5d666c8-3d64-4297-89eb-07c7e0a316e6-config-volume\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.287962 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5d666c8-3d64-4297-89eb-07c7e0a316e6-secret-volume\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.309913 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b5h4\" (UniqueName: \"kubernetes.io/projected/f5d666c8-3d64-4297-89eb-07c7e0a316e6-kube-api-access-4b5h4\") pod \"collect-profiles-29323200-ppgrz\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.476016 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:00 crc kubenswrapper[4960]: I1002 08:00:00.773600 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz"] Oct 02 08:00:01 crc kubenswrapper[4960]: I1002 08:00:01.511846 4960 generic.go:334] "Generic (PLEG): container finished" podID="f5d666c8-3d64-4297-89eb-07c7e0a316e6" containerID="14041f654a74645cca51e60a458eaef5fa73bbf411b3e005c22100b4e66c89dc" exitCode=0 Oct 02 08:00:01 crc kubenswrapper[4960]: I1002 08:00:01.511919 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" event={"ID":"f5d666c8-3d64-4297-89eb-07c7e0a316e6","Type":"ContainerDied","Data":"14041f654a74645cca51e60a458eaef5fa73bbf411b3e005c22100b4e66c89dc"} Oct 02 08:00:01 crc kubenswrapper[4960]: I1002 08:00:01.512451 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" event={"ID":"f5d666c8-3d64-4297-89eb-07c7e0a316e6","Type":"ContainerStarted","Data":"ee6e36437b4ba53e16a0c97664f32f3a1a8415f860c24b7a48bcfb4e957ac338"} Oct 02 08:00:02 crc kubenswrapper[4960]: I1002 08:00:02.874620 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:02 crc kubenswrapper[4960]: I1002 08:00:02.936379 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5d666c8-3d64-4297-89eb-07c7e0a316e6-secret-volume\") pod \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " Oct 02 08:00:02 crc kubenswrapper[4960]: I1002 08:00:02.936734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5d666c8-3d64-4297-89eb-07c7e0a316e6-config-volume\") pod \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " Oct 02 08:00:02 crc kubenswrapper[4960]: I1002 08:00:02.936823 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b5h4\" (UniqueName: \"kubernetes.io/projected/f5d666c8-3d64-4297-89eb-07c7e0a316e6-kube-api-access-4b5h4\") pod \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\" (UID: \"f5d666c8-3d64-4297-89eb-07c7e0a316e6\") " Oct 02 08:00:02 crc kubenswrapper[4960]: I1002 08:00:02.937695 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5d666c8-3d64-4297-89eb-07c7e0a316e6-config-volume" (OuterVolumeSpecName: "config-volume") pod "f5d666c8-3d64-4297-89eb-07c7e0a316e6" (UID: "f5d666c8-3d64-4297-89eb-07c7e0a316e6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:00:02 crc kubenswrapper[4960]: I1002 08:00:02.945494 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d666c8-3d64-4297-89eb-07c7e0a316e6-kube-api-access-4b5h4" (OuterVolumeSpecName: "kube-api-access-4b5h4") pod "f5d666c8-3d64-4297-89eb-07c7e0a316e6" (UID: "f5d666c8-3d64-4297-89eb-07c7e0a316e6"). InnerVolumeSpecName "kube-api-access-4b5h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:00:02 crc kubenswrapper[4960]: I1002 08:00:02.949316 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d666c8-3d64-4297-89eb-07c7e0a316e6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f5d666c8-3d64-4297-89eb-07c7e0a316e6" (UID: "f5d666c8-3d64-4297-89eb-07c7e0a316e6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.038594 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5d666c8-3d64-4297-89eb-07c7e0a316e6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.038635 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b5h4\" (UniqueName: \"kubernetes.io/projected/f5d666c8-3d64-4297-89eb-07c7e0a316e6-kube-api-access-4b5h4\") on node \"crc\" DevicePath \"\"" Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.038648 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5d666c8-3d64-4297-89eb-07c7e0a316e6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.530459 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" event={"ID":"f5d666c8-3d64-4297-89eb-07c7e0a316e6","Type":"ContainerDied","Data":"ee6e36437b4ba53e16a0c97664f32f3a1a8415f860c24b7a48bcfb4e957ac338"} Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.530520 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee6e36437b4ba53e16a0c97664f32f3a1a8415f860c24b7a48bcfb4e957ac338" Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.530540 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz" Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.964110 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn"] Oct 02 08:00:03 crc kubenswrapper[4960]: I1002 08:00:03.973338 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g6jvn"] Oct 02 08:00:04 crc kubenswrapper[4960]: I1002 08:00:04.352185 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a0f863-045c-4251-865d-327d1c1397e5" path="/var/lib/kubelet/pods/97a0f863-045c-4251-865d-327d1c1397e5/volumes" Oct 02 08:00:45 crc kubenswrapper[4960]: I1002 08:00:45.093797 4960 scope.go:117] "RemoveContainer" containerID="be069d06de2e3d9000a4f6cd982946be60106145e944ec354eed034347a8211c" Oct 02 08:00:59 crc kubenswrapper[4960]: I1002 08:00:59.150705 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:00:59 crc kubenswrapper[4960]: I1002 08:00:59.151491 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.158337 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323201-k4bsb"] Oct 02 08:01:00 crc kubenswrapper[4960]: E1002 08:01:00.158807 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d666c8-3d64-4297-89eb-07c7e0a316e6" containerName="collect-profiles" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.158821 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d666c8-3d64-4297-89eb-07c7e0a316e6" containerName="collect-profiles" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.159018 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d666c8-3d64-4297-89eb-07c7e0a316e6" containerName="collect-profiles" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.159749 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.175126 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323201-k4bsb"] Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.268524 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-fernet-keys\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.268626 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-config-data\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.268675 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzbxv\" (UniqueName: \"kubernetes.io/projected/20148c75-2594-4fa9-b38e-1fdc97757c0f-kube-api-access-mzbxv\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.268760 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-combined-ca-bundle\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.371117 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-fernet-keys\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.371384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-config-data\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.371466 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzbxv\" (UniqueName: \"kubernetes.io/projected/20148c75-2594-4fa9-b38e-1fdc97757c0f-kube-api-access-mzbxv\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.371611 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-combined-ca-bundle\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.379837 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-config-data\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.380355 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-combined-ca-bundle\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.383801 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-fernet-keys\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.398540 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzbxv\" (UniqueName: \"kubernetes.io/projected/20148c75-2594-4fa9-b38e-1fdc97757c0f-kube-api-access-mzbxv\") pod \"keystone-cron-29323201-k4bsb\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.484578 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:00 crc kubenswrapper[4960]: I1002 08:01:00.922944 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323201-k4bsb"] Oct 02 08:01:01 crc kubenswrapper[4960]: I1002 08:01:01.163684 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-k4bsb" event={"ID":"20148c75-2594-4fa9-b38e-1fdc97757c0f","Type":"ContainerStarted","Data":"a589d97cba864e3457be15ca6c9d0bb2e92011eaa107eb019d482de1121695c9"} Oct 02 08:01:01 crc kubenswrapper[4960]: I1002 08:01:01.164389 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-k4bsb" event={"ID":"20148c75-2594-4fa9-b38e-1fdc97757c0f","Type":"ContainerStarted","Data":"827f581077698a809c50286455c1209c7d0f1a6abd19be12426e571f2bb57073"} Oct 02 08:01:01 crc kubenswrapper[4960]: I1002 08:01:01.185484 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323201-k4bsb" podStartSLOduration=1.185431019 podStartE2EDuration="1.185431019s" podCreationTimestamp="2025-10-02 08:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:01:01.18061259 +0000 UTC m=+2682.212558877" watchObservedRunningTime="2025-10-02 08:01:01.185431019 +0000 UTC m=+2682.217377306" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.089742 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xlf97"] Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.092452 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.109219 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlf97"] Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.156389 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-catalog-content\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.156459 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nqh9\" (UniqueName: \"kubernetes.io/projected/ae037dfd-e05b-4422-bbc4-9f8842284dc7-kube-api-access-5nqh9\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.156529 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-utilities\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.185300 4960 generic.go:334] "Generic (PLEG): container finished" podID="20148c75-2594-4fa9-b38e-1fdc97757c0f" containerID="a589d97cba864e3457be15ca6c9d0bb2e92011eaa107eb019d482de1121695c9" exitCode=0 Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.185355 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-k4bsb" event={"ID":"20148c75-2594-4fa9-b38e-1fdc97757c0f","Type":"ContainerDied","Data":"a589d97cba864e3457be15ca6c9d0bb2e92011eaa107eb019d482de1121695c9"} Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.258180 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-utilities\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.258668 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-catalog-content\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.258809 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-utilities\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.259014 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nqh9\" (UniqueName: \"kubernetes.io/projected/ae037dfd-e05b-4422-bbc4-9f8842284dc7-kube-api-access-5nqh9\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.259222 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-catalog-content\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.280137 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nqh9\" (UniqueName: \"kubernetes.io/projected/ae037dfd-e05b-4422-bbc4-9f8842284dc7-kube-api-access-5nqh9\") pod \"redhat-marketplace-xlf97\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.427720 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:03 crc kubenswrapper[4960]: I1002 08:01:03.918643 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlf97"] Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.196586 4960 generic.go:334] "Generic (PLEG): container finished" podID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerID="82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58" exitCode=0 Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.196646 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlf97" event={"ID":"ae037dfd-e05b-4422-bbc4-9f8842284dc7","Type":"ContainerDied","Data":"82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58"} Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.196719 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlf97" event={"ID":"ae037dfd-e05b-4422-bbc4-9f8842284dc7","Type":"ContainerStarted","Data":"8508f05aa0b004591e616266bfb8272f28fe153ee26048edd2025b85cad45ed6"} Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.200194 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.617704 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.800104 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-config-data\") pod \"20148c75-2594-4fa9-b38e-1fdc97757c0f\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.800302 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-fernet-keys\") pod \"20148c75-2594-4fa9-b38e-1fdc97757c0f\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.800468 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzbxv\" (UniqueName: \"kubernetes.io/projected/20148c75-2594-4fa9-b38e-1fdc97757c0f-kube-api-access-mzbxv\") pod \"20148c75-2594-4fa9-b38e-1fdc97757c0f\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.800512 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-combined-ca-bundle\") pod \"20148c75-2594-4fa9-b38e-1fdc97757c0f\" (UID: \"20148c75-2594-4fa9-b38e-1fdc97757c0f\") " Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.807993 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20148c75-2594-4fa9-b38e-1fdc97757c0f-kube-api-access-mzbxv" (OuterVolumeSpecName: "kube-api-access-mzbxv") pod "20148c75-2594-4fa9-b38e-1fdc97757c0f" (UID: "20148c75-2594-4fa9-b38e-1fdc97757c0f"). InnerVolumeSpecName "kube-api-access-mzbxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.814219 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "20148c75-2594-4fa9-b38e-1fdc97757c0f" (UID: "20148c75-2594-4fa9-b38e-1fdc97757c0f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.830061 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20148c75-2594-4fa9-b38e-1fdc97757c0f" (UID: "20148c75-2594-4fa9-b38e-1fdc97757c0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.855539 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-config-data" (OuterVolumeSpecName: "config-data") pod "20148c75-2594-4fa9-b38e-1fdc97757c0f" (UID: "20148c75-2594-4fa9-b38e-1fdc97757c0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.902873 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.902906 4960 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.902916 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzbxv\" (UniqueName: \"kubernetes.io/projected/20148c75-2594-4fa9-b38e-1fdc97757c0f-kube-api-access-mzbxv\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:04 crc kubenswrapper[4960]: I1002 08:01:04.902924 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20148c75-2594-4fa9-b38e-1fdc97757c0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:05 crc kubenswrapper[4960]: I1002 08:01:05.228635 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-k4bsb" event={"ID":"20148c75-2594-4fa9-b38e-1fdc97757c0f","Type":"ContainerDied","Data":"827f581077698a809c50286455c1209c7d0f1a6abd19be12426e571f2bb57073"} Oct 02 08:01:05 crc kubenswrapper[4960]: I1002 08:01:05.228774 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="827f581077698a809c50286455c1209c7d0f1a6abd19be12426e571f2bb57073" Oct 02 08:01:05 crc kubenswrapper[4960]: I1002 08:01:05.228694 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-k4bsb" Oct 02 08:01:05 crc kubenswrapper[4960]: I1002 08:01:05.240992 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlf97" event={"ID":"ae037dfd-e05b-4422-bbc4-9f8842284dc7","Type":"ContainerStarted","Data":"134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5"} Oct 02 08:01:06 crc kubenswrapper[4960]: I1002 08:01:06.251913 4960 generic.go:334] "Generic (PLEG): container finished" podID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerID="134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5" exitCode=0 Oct 02 08:01:06 crc kubenswrapper[4960]: I1002 08:01:06.252103 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlf97" event={"ID":"ae037dfd-e05b-4422-bbc4-9f8842284dc7","Type":"ContainerDied","Data":"134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5"} Oct 02 08:01:07 crc kubenswrapper[4960]: I1002 08:01:07.269732 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlf97" event={"ID":"ae037dfd-e05b-4422-bbc4-9f8842284dc7","Type":"ContainerStarted","Data":"252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62"} Oct 02 08:01:07 crc kubenswrapper[4960]: I1002 08:01:07.298268 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xlf97" podStartSLOduration=1.7508286960000001 podStartE2EDuration="4.298251349s" podCreationTimestamp="2025-10-02 08:01:03 +0000 UTC" firstStartedPulling="2025-10-02 08:01:04.199700413 +0000 UTC m=+2685.231646730" lastFinishedPulling="2025-10-02 08:01:06.747123056 +0000 UTC m=+2687.779069383" observedRunningTime="2025-10-02 08:01:07.290216823 +0000 UTC m=+2688.322163110" watchObservedRunningTime="2025-10-02 08:01:07.298251349 +0000 UTC m=+2688.330197626" Oct 02 08:01:13 crc kubenswrapper[4960]: I1002 08:01:13.429641 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:13 crc kubenswrapper[4960]: I1002 08:01:13.430477 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:13 crc kubenswrapper[4960]: I1002 08:01:13.497346 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:14 crc kubenswrapper[4960]: I1002 08:01:14.401438 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:14 crc kubenswrapper[4960]: I1002 08:01:14.459792 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlf97"] Oct 02 08:01:16 crc kubenswrapper[4960]: I1002 08:01:16.385489 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xlf97" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="registry-server" containerID="cri-o://252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62" gracePeriod=2 Oct 02 08:01:16 crc kubenswrapper[4960]: I1002 08:01:16.921602 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.046075 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nqh9\" (UniqueName: \"kubernetes.io/projected/ae037dfd-e05b-4422-bbc4-9f8842284dc7-kube-api-access-5nqh9\") pod \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.046221 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-catalog-content\") pod \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.046315 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-utilities\") pod \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\" (UID: \"ae037dfd-e05b-4422-bbc4-9f8842284dc7\") " Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.047324 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-utilities" (OuterVolumeSpecName: "utilities") pod "ae037dfd-e05b-4422-bbc4-9f8842284dc7" (UID: "ae037dfd-e05b-4422-bbc4-9f8842284dc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.052307 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae037dfd-e05b-4422-bbc4-9f8842284dc7-kube-api-access-5nqh9" (OuterVolumeSpecName: "kube-api-access-5nqh9") pod "ae037dfd-e05b-4422-bbc4-9f8842284dc7" (UID: "ae037dfd-e05b-4422-bbc4-9f8842284dc7"). InnerVolumeSpecName "kube-api-access-5nqh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.059883 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae037dfd-e05b-4422-bbc4-9f8842284dc7" (UID: "ae037dfd-e05b-4422-bbc4-9f8842284dc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.148533 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nqh9\" (UniqueName: \"kubernetes.io/projected/ae037dfd-e05b-4422-bbc4-9f8842284dc7-kube-api-access-5nqh9\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.148578 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.148590 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae037dfd-e05b-4422-bbc4-9f8842284dc7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.400441 4960 generic.go:334] "Generic (PLEG): container finished" podID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerID="252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62" exitCode=0 Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.400502 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlf97" event={"ID":"ae037dfd-e05b-4422-bbc4-9f8842284dc7","Type":"ContainerDied","Data":"252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62"} Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.400584 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlf97" event={"ID":"ae037dfd-e05b-4422-bbc4-9f8842284dc7","Type":"ContainerDied","Data":"8508f05aa0b004591e616266bfb8272f28fe153ee26048edd2025b85cad45ed6"} Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.400580 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlf97" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.400609 4960 scope.go:117] "RemoveContainer" containerID="252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.426226 4960 scope.go:117] "RemoveContainer" containerID="134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.504239 4960 scope.go:117] "RemoveContainer" containerID="82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.528931 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlf97"] Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.535738 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlf97"] Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.541758 4960 scope.go:117] "RemoveContainer" containerID="252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62" Oct 02 08:01:17 crc kubenswrapper[4960]: E1002 08:01:17.543118 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62\": container with ID starting with 252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62 not found: ID does not exist" containerID="252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.543170 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62"} err="failed to get container status \"252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62\": rpc error: code = NotFound desc = could not find container \"252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62\": container with ID starting with 252a43eaa1a496da4be58013fe92cfa67d363126b81f5bd75b8a43f08e811e62 not found: ID does not exist" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.543204 4960 scope.go:117] "RemoveContainer" containerID="134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5" Oct 02 08:01:17 crc kubenswrapper[4960]: E1002 08:01:17.543622 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5\": container with ID starting with 134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5 not found: ID does not exist" containerID="134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.543659 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5"} err="failed to get container status \"134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5\": rpc error: code = NotFound desc = could not find container \"134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5\": container with ID starting with 134d91512a861750147a02ff680a1908d0f5229e4d33c74bb085151576a396e5 not found: ID does not exist" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.544033 4960 scope.go:117] "RemoveContainer" containerID="82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58" Oct 02 08:01:17 crc kubenswrapper[4960]: E1002 08:01:17.544722 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58\": container with ID starting with 82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58 not found: ID does not exist" containerID="82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58" Oct 02 08:01:17 crc kubenswrapper[4960]: I1002 08:01:17.544813 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58"} err="failed to get container status \"82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58\": rpc error: code = NotFound desc = could not find container \"82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58\": container with ID starting with 82232c00aa1c342a15e40191b0ab1dbd20505771b2fd80be0b71197129c9ea58 not found: ID does not exist" Oct 02 08:01:18 crc kubenswrapper[4960]: I1002 08:01:18.347712 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" path="/var/lib/kubelet/pods/ae037dfd-e05b-4422-bbc4-9f8842284dc7/volumes" Oct 02 08:01:29 crc kubenswrapper[4960]: I1002 08:01:29.150562 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:01:29 crc kubenswrapper[4960]: I1002 08:01:29.151658 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.150251 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.151214 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.151309 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.152679 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"697e5417554915e415b18e62ccd2b91fc62b5307191461e4f59ddcb5fd447d6d"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.152809 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://697e5417554915e415b18e62ccd2b91fc62b5307191461e4f59ddcb5fd447d6d" gracePeriod=600 Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.883354 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="697e5417554915e415b18e62ccd2b91fc62b5307191461e4f59ddcb5fd447d6d" exitCode=0 Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.883491 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"697e5417554915e415b18e62ccd2b91fc62b5307191461e4f59ddcb5fd447d6d"} Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.884054 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498"} Oct 02 08:01:59 crc kubenswrapper[4960]: I1002 08:01:59.884093 4960 scope.go:117] "RemoveContainer" containerID="f93e480ed6343e93f1b9d6f47b5cf910d58a2a34c7cd8117c41d3a32bb35c755" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.724015 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-phz28"] Oct 02 08:02:05 crc kubenswrapper[4960]: E1002 08:02:05.733644 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="registry-server" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.733849 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="registry-server" Oct 02 08:02:05 crc kubenswrapper[4960]: E1002 08:02:05.734075 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20148c75-2594-4fa9-b38e-1fdc97757c0f" containerName="keystone-cron" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.734216 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="20148c75-2594-4fa9-b38e-1fdc97757c0f" containerName="keystone-cron" Oct 02 08:02:05 crc kubenswrapper[4960]: E1002 08:02:05.734359 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="extract-utilities" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.734498 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="extract-utilities" Oct 02 08:02:05 crc kubenswrapper[4960]: E1002 08:02:05.734646 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="extract-content" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.734778 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="extract-content" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.735281 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae037dfd-e05b-4422-bbc4-9f8842284dc7" containerName="registry-server" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.735475 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="20148c75-2594-4fa9-b38e-1fdc97757c0f" containerName="keystone-cron" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.738193 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-phz28"] Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.738387 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.832864 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-utilities\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.833025 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-catalog-content\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.833107 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wnpl\" (UniqueName: \"kubernetes.io/projected/c820a4d8-587b-45b4-a746-314be19e4a2a-kube-api-access-5wnpl\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.934945 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-utilities\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.935089 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-catalog-content\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.935143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wnpl\" (UniqueName: \"kubernetes.io/projected/c820a4d8-587b-45b4-a746-314be19e4a2a-kube-api-access-5wnpl\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.935856 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-catalog-content\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.937108 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-utilities\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:05 crc kubenswrapper[4960]: I1002 08:02:05.975490 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wnpl\" (UniqueName: \"kubernetes.io/projected/c820a4d8-587b-45b4-a746-314be19e4a2a-kube-api-access-5wnpl\") pod \"certified-operators-phz28\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:06 crc kubenswrapper[4960]: I1002 08:02:06.079105 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:06 crc kubenswrapper[4960]: I1002 08:02:06.575480 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-phz28"] Oct 02 08:02:06 crc kubenswrapper[4960]: I1002 08:02:06.979842 4960 generic.go:334] "Generic (PLEG): container finished" podID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerID="5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32" exitCode=0 Oct 02 08:02:06 crc kubenswrapper[4960]: I1002 08:02:06.979942 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-phz28" event={"ID":"c820a4d8-587b-45b4-a746-314be19e4a2a","Type":"ContainerDied","Data":"5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32"} Oct 02 08:02:06 crc kubenswrapper[4960]: I1002 08:02:06.980594 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-phz28" event={"ID":"c820a4d8-587b-45b4-a746-314be19e4a2a","Type":"ContainerStarted","Data":"9e1033e3b368c4badff2192abc90a771b75aed58734a640ad053250db1148ca3"} Oct 02 08:02:07 crc kubenswrapper[4960]: I1002 08:02:07.996428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-phz28" event={"ID":"c820a4d8-587b-45b4-a746-314be19e4a2a","Type":"ContainerStarted","Data":"79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69"} Oct 02 08:02:09 crc kubenswrapper[4960]: I1002 08:02:09.014263 4960 generic.go:334] "Generic (PLEG): container finished" podID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerID="79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69" exitCode=0 Oct 02 08:02:09 crc kubenswrapper[4960]: I1002 08:02:09.014363 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-phz28" event={"ID":"c820a4d8-587b-45b4-a746-314be19e4a2a","Type":"ContainerDied","Data":"79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69"} Oct 02 08:02:10 crc kubenswrapper[4960]: I1002 08:02:10.030423 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-phz28" event={"ID":"c820a4d8-587b-45b4-a746-314be19e4a2a","Type":"ContainerStarted","Data":"f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399"} Oct 02 08:02:10 crc kubenswrapper[4960]: I1002 08:02:10.077827 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-phz28" podStartSLOduration=2.410111832 podStartE2EDuration="5.077791134s" podCreationTimestamp="2025-10-02 08:02:05 +0000 UTC" firstStartedPulling="2025-10-02 08:02:06.982824614 +0000 UTC m=+2748.014770901" lastFinishedPulling="2025-10-02 08:02:09.650503916 +0000 UTC m=+2750.682450203" observedRunningTime="2025-10-02 08:02:10.063522141 +0000 UTC m=+2751.095468448" watchObservedRunningTime="2025-10-02 08:02:10.077791134 +0000 UTC m=+2751.109737421" Oct 02 08:02:16 crc kubenswrapper[4960]: I1002 08:02:16.080158 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:16 crc kubenswrapper[4960]: I1002 08:02:16.081342 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:16 crc kubenswrapper[4960]: I1002 08:02:16.162150 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:16 crc kubenswrapper[4960]: I1002 08:02:16.247084 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:16 crc kubenswrapper[4960]: I1002 08:02:16.419894 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-phz28"] Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.145899 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-phz28" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="registry-server" containerID="cri-o://f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399" gracePeriod=2 Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.695512 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.881299 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-catalog-content\") pod \"c820a4d8-587b-45b4-a746-314be19e4a2a\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.881633 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wnpl\" (UniqueName: \"kubernetes.io/projected/c820a4d8-587b-45b4-a746-314be19e4a2a-kube-api-access-5wnpl\") pod \"c820a4d8-587b-45b4-a746-314be19e4a2a\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.881680 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-utilities\") pod \"c820a4d8-587b-45b4-a746-314be19e4a2a\" (UID: \"c820a4d8-587b-45b4-a746-314be19e4a2a\") " Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.882570 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-utilities" (OuterVolumeSpecName: "utilities") pod "c820a4d8-587b-45b4-a746-314be19e4a2a" (UID: "c820a4d8-587b-45b4-a746-314be19e4a2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.892728 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c820a4d8-587b-45b4-a746-314be19e4a2a-kube-api-access-5wnpl" (OuterVolumeSpecName: "kube-api-access-5wnpl") pod "c820a4d8-587b-45b4-a746-314be19e4a2a" (UID: "c820a4d8-587b-45b4-a746-314be19e4a2a"). InnerVolumeSpecName "kube-api-access-5wnpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.941099 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c820a4d8-587b-45b4-a746-314be19e4a2a" (UID: "c820a4d8-587b-45b4-a746-314be19e4a2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.984108 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.984140 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wnpl\" (UniqueName: \"kubernetes.io/projected/c820a4d8-587b-45b4-a746-314be19e4a2a-kube-api-access-5wnpl\") on node \"crc\" DevicePath \"\"" Oct 02 08:02:18 crc kubenswrapper[4960]: I1002 08:02:18.984155 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c820a4d8-587b-45b4-a746-314be19e4a2a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.159785 4960 generic.go:334] "Generic (PLEG): container finished" podID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerID="f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399" exitCode=0 Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.159846 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-phz28" event={"ID":"c820a4d8-587b-45b4-a746-314be19e4a2a","Type":"ContainerDied","Data":"f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399"} Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.159882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-phz28" event={"ID":"c820a4d8-587b-45b4-a746-314be19e4a2a","Type":"ContainerDied","Data":"9e1033e3b368c4badff2192abc90a771b75aed58734a640ad053250db1148ca3"} Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.159907 4960 scope.go:117] "RemoveContainer" containerID="f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.159922 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-phz28" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.199110 4960 scope.go:117] "RemoveContainer" containerID="79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.205968 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-phz28"] Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.212528 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-phz28"] Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.226461 4960 scope.go:117] "RemoveContainer" containerID="5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.273770 4960 scope.go:117] "RemoveContainer" containerID="f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399" Oct 02 08:02:19 crc kubenswrapper[4960]: E1002 08:02:19.274452 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399\": container with ID starting with f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399 not found: ID does not exist" containerID="f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.274520 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399"} err="failed to get container status \"f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399\": rpc error: code = NotFound desc = could not find container \"f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399\": container with ID starting with f0ed4f044b2d66212490313ebf16afaa98f494a8ab4472fe3a8e0b6abb87c399 not found: ID does not exist" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.274573 4960 scope.go:117] "RemoveContainer" containerID="79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69" Oct 02 08:02:19 crc kubenswrapper[4960]: E1002 08:02:19.274956 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69\": container with ID starting with 79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69 not found: ID does not exist" containerID="79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.275011 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69"} err="failed to get container status \"79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69\": rpc error: code = NotFound desc = could not find container \"79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69\": container with ID starting with 79777e0e0403f20797dd2f0d5ea8c51158451dbcf27f05ddb73606f96b107b69 not found: ID does not exist" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.275040 4960 scope.go:117] "RemoveContainer" containerID="5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32" Oct 02 08:02:19 crc kubenswrapper[4960]: E1002 08:02:19.275444 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32\": container with ID starting with 5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32 not found: ID does not exist" containerID="5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32" Oct 02 08:02:19 crc kubenswrapper[4960]: I1002 08:02:19.275481 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32"} err="failed to get container status \"5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32\": rpc error: code = NotFound desc = could not find container \"5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32\": container with ID starting with 5e4e475101b55f5f9dafb06f4391799410845769a106aef102793495bfa7ff32 not found: ID does not exist" Oct 02 08:02:20 crc kubenswrapper[4960]: I1002 08:02:20.344846 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" path="/var/lib/kubelet/pods/c820a4d8-587b-45b4-a746-314be19e4a2a/volumes" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.743144 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-68bzn"] Oct 02 08:03:34 crc kubenswrapper[4960]: E1002 08:03:34.744664 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="registry-server" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.744688 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="registry-server" Oct 02 08:03:34 crc kubenswrapper[4960]: E1002 08:03:34.744715 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="extract-utilities" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.744727 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="extract-utilities" Oct 02 08:03:34 crc kubenswrapper[4960]: E1002 08:03:34.744750 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="extract-content" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.744762 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="extract-content" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.745171 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c820a4d8-587b-45b4-a746-314be19e4a2a" containerName="registry-server" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.747331 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.760712 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68bzn"] Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.888800 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xcq8\" (UniqueName: \"kubernetes.io/projected/25873846-5b45-4500-8e66-270f6fdc2722-kube-api-access-7xcq8\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.889098 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-utilities\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.889156 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-catalog-content\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.991909 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-utilities\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.992026 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-catalog-content\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.992123 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xcq8\" (UniqueName: \"kubernetes.io/projected/25873846-5b45-4500-8e66-270f6fdc2722-kube-api-access-7xcq8\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.992649 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-utilities\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:34 crc kubenswrapper[4960]: I1002 08:03:34.992651 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-catalog-content\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:35 crc kubenswrapper[4960]: I1002 08:03:35.027572 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xcq8\" (UniqueName: \"kubernetes.io/projected/25873846-5b45-4500-8e66-270f6fdc2722-kube-api-access-7xcq8\") pod \"redhat-operators-68bzn\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:35 crc kubenswrapper[4960]: I1002 08:03:35.085930 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:35 crc kubenswrapper[4960]: I1002 08:03:35.635920 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68bzn"] Oct 02 08:03:36 crc kubenswrapper[4960]: I1002 08:03:36.125180 4960 generic.go:334] "Generic (PLEG): container finished" podID="25873846-5b45-4500-8e66-270f6fdc2722" containerID="809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af" exitCode=0 Oct 02 08:03:36 crc kubenswrapper[4960]: I1002 08:03:36.125281 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68bzn" event={"ID":"25873846-5b45-4500-8e66-270f6fdc2722","Type":"ContainerDied","Data":"809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af"} Oct 02 08:03:36 crc kubenswrapper[4960]: I1002 08:03:36.125639 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68bzn" event={"ID":"25873846-5b45-4500-8e66-270f6fdc2722","Type":"ContainerStarted","Data":"73e3b8f02ca848f6a1800ac16e27988e5981bd53e456d6641bacbb20f600abba"} Oct 02 08:03:36 crc kubenswrapper[4960]: I1002 08:03:36.942359 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zkdjz"] Oct 02 08:03:36 crc kubenswrapper[4960]: I1002 08:03:36.945384 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:36 crc kubenswrapper[4960]: I1002 08:03:36.950202 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zkdjz"] Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.040245 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj6wg\" (UniqueName: \"kubernetes.io/projected/b58338a1-91dc-4e64-b257-33d487e83d7f-kube-api-access-kj6wg\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.040760 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-utilities\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.040888 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-catalog-content\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.139319 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68bzn" event={"ID":"25873846-5b45-4500-8e66-270f6fdc2722","Type":"ContainerStarted","Data":"c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338"} Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.142913 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-catalog-content\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.143034 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj6wg\" (UniqueName: \"kubernetes.io/projected/b58338a1-91dc-4e64-b257-33d487e83d7f-kube-api-access-kj6wg\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.143080 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-utilities\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.143710 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-utilities\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.143739 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-catalog-content\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.169960 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj6wg\" (UniqueName: \"kubernetes.io/projected/b58338a1-91dc-4e64-b257-33d487e83d7f-kube-api-access-kj6wg\") pod \"community-operators-zkdjz\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.334164 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:37 crc kubenswrapper[4960]: I1002 08:03:37.836205 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zkdjz"] Oct 02 08:03:37 crc kubenswrapper[4960]: W1002 08:03:37.846022 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb58338a1_91dc_4e64_b257_33d487e83d7f.slice/crio-edf3103d12a7b433dee3b08a680af1378eb624ce4e33f59e25a3cb2a51d59eb6 WatchSource:0}: Error finding container edf3103d12a7b433dee3b08a680af1378eb624ce4e33f59e25a3cb2a51d59eb6: Status 404 returned error can't find the container with id edf3103d12a7b433dee3b08a680af1378eb624ce4e33f59e25a3cb2a51d59eb6 Oct 02 08:03:38 crc kubenswrapper[4960]: I1002 08:03:38.150930 4960 generic.go:334] "Generic (PLEG): container finished" podID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerID="6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627" exitCode=0 Oct 02 08:03:38 crc kubenswrapper[4960]: I1002 08:03:38.151052 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkdjz" event={"ID":"b58338a1-91dc-4e64-b257-33d487e83d7f","Type":"ContainerDied","Data":"6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627"} Oct 02 08:03:38 crc kubenswrapper[4960]: I1002 08:03:38.151151 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkdjz" event={"ID":"b58338a1-91dc-4e64-b257-33d487e83d7f","Type":"ContainerStarted","Data":"edf3103d12a7b433dee3b08a680af1378eb624ce4e33f59e25a3cb2a51d59eb6"} Oct 02 08:03:38 crc kubenswrapper[4960]: I1002 08:03:38.154764 4960 generic.go:334] "Generic (PLEG): container finished" podID="25873846-5b45-4500-8e66-270f6fdc2722" containerID="c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338" exitCode=0 Oct 02 08:03:38 crc kubenswrapper[4960]: I1002 08:03:38.154886 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68bzn" event={"ID":"25873846-5b45-4500-8e66-270f6fdc2722","Type":"ContainerDied","Data":"c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338"} Oct 02 08:03:39 crc kubenswrapper[4960]: I1002 08:03:39.172571 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkdjz" event={"ID":"b58338a1-91dc-4e64-b257-33d487e83d7f","Type":"ContainerStarted","Data":"2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7"} Oct 02 08:03:39 crc kubenswrapper[4960]: I1002 08:03:39.177689 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68bzn" event={"ID":"25873846-5b45-4500-8e66-270f6fdc2722","Type":"ContainerStarted","Data":"17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00"} Oct 02 08:03:39 crc kubenswrapper[4960]: I1002 08:03:39.222521 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-68bzn" podStartSLOduration=2.6769372860000002 podStartE2EDuration="5.222497508s" podCreationTimestamp="2025-10-02 08:03:34 +0000 UTC" firstStartedPulling="2025-10-02 08:03:36.127218953 +0000 UTC m=+2837.159165240" lastFinishedPulling="2025-10-02 08:03:38.672779175 +0000 UTC m=+2839.704725462" observedRunningTime="2025-10-02 08:03:39.21376497 +0000 UTC m=+2840.245711267" watchObservedRunningTime="2025-10-02 08:03:39.222497508 +0000 UTC m=+2840.254443805" Oct 02 08:03:40 crc kubenswrapper[4960]: I1002 08:03:40.190566 4960 generic.go:334] "Generic (PLEG): container finished" podID="f2f4ac6d-fe72-4957-bcdb-983b63fa172b" containerID="db5f455cf4a4449f523f2053ba51636bced5c48cc1b015aedd7d3cc50a6aa0ab" exitCode=0 Oct 02 08:03:40 crc kubenswrapper[4960]: I1002 08:03:40.190641 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" event={"ID":"f2f4ac6d-fe72-4957-bcdb-983b63fa172b","Type":"ContainerDied","Data":"db5f455cf4a4449f523f2053ba51636bced5c48cc1b015aedd7d3cc50a6aa0ab"} Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.728769 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.863327 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg4jx\" (UniqueName: \"kubernetes.io/projected/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-kube-api-access-sg4jx\") pod \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.863549 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-secret-0\") pod \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.863611 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ssh-key\") pod \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.863661 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ceph\") pod \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.863695 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-combined-ca-bundle\") pod \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.863753 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-inventory\") pod \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\" (UID: \"f2f4ac6d-fe72-4957-bcdb-983b63fa172b\") " Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.874676 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-kube-api-access-sg4jx" (OuterVolumeSpecName: "kube-api-access-sg4jx") pod "f2f4ac6d-fe72-4957-bcdb-983b63fa172b" (UID: "f2f4ac6d-fe72-4957-bcdb-983b63fa172b"). InnerVolumeSpecName "kube-api-access-sg4jx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.880066 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ceph" (OuterVolumeSpecName: "ceph") pod "f2f4ac6d-fe72-4957-bcdb-983b63fa172b" (UID: "f2f4ac6d-fe72-4957-bcdb-983b63fa172b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.890834 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f2f4ac6d-fe72-4957-bcdb-983b63fa172b" (UID: "f2f4ac6d-fe72-4957-bcdb-983b63fa172b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.908260 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "f2f4ac6d-fe72-4957-bcdb-983b63fa172b" (UID: "f2f4ac6d-fe72-4957-bcdb-983b63fa172b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.908316 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-inventory" (OuterVolumeSpecName: "inventory") pod "f2f4ac6d-fe72-4957-bcdb-983b63fa172b" (UID: "f2f4ac6d-fe72-4957-bcdb-983b63fa172b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.911506 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f2f4ac6d-fe72-4957-bcdb-983b63fa172b" (UID: "f2f4ac6d-fe72-4957-bcdb-983b63fa172b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.967048 4960 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.967096 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.967108 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg4jx\" (UniqueName: \"kubernetes.io/projected/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-kube-api-access-sg4jx\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.967119 4960 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.967129 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:41 crc kubenswrapper[4960]: I1002 08:03:41.967168 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2f4ac6d-fe72-4957-bcdb-983b63fa172b-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.211366 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" event={"ID":"f2f4ac6d-fe72-4957-bcdb-983b63fa172b","Type":"ContainerDied","Data":"ff060e1d68c875fadb684cf148c478866afb54332f2a7db1c6252d88157fa998"} Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.211451 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff060e1d68c875fadb684cf148c478866afb54332f2a7db1c6252d88157fa998" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.211399 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.214070 4960 generic.go:334] "Generic (PLEG): container finished" podID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerID="2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7" exitCode=0 Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.214118 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkdjz" event={"ID":"b58338a1-91dc-4e64-b257-33d487e83d7f","Type":"ContainerDied","Data":"2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7"} Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.314087 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z"] Oct 02 08:03:42 crc kubenswrapper[4960]: E1002 08:03:42.321117 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f4ac6d-fe72-4957-bcdb-983b63fa172b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.321166 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f4ac6d-fe72-4957-bcdb-983b63fa172b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.321649 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2f4ac6d-fe72-4957-bcdb-983b63fa172b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.322540 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.326077 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z"] Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.327959 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xpr8s" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.328188 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.328313 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.329559 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.329686 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.329792 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.329920 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.330144 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.330275 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.477482 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.477552 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmr6t\" (UniqueName: \"kubernetes.io/projected/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-kube-api-access-qmr6t\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.477780 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.477888 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.477937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.478011 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.478056 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.478090 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.478114 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.478311 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.478359 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581030 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581106 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581137 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581185 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581210 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581233 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581272 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.581296 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.582547 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.582604 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmr6t\" (UniqueName: \"kubernetes.io/projected/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-kube-api-access-qmr6t\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.582648 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.582810 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.583323 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.585949 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.586841 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.587155 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.587764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.588026 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.591356 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.592490 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.592515 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.616658 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmr6t\" (UniqueName: \"kubernetes.io/projected/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-kube-api-access-qmr6t\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:42 crc kubenswrapper[4960]: I1002 08:03:42.649564 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:03:43 crc kubenswrapper[4960]: I1002 08:03:43.227242 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkdjz" event={"ID":"b58338a1-91dc-4e64-b257-33d487e83d7f","Type":"ContainerStarted","Data":"174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c"} Oct 02 08:03:43 crc kubenswrapper[4960]: I1002 08:03:43.257642 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zkdjz" podStartSLOduration=2.75597519 podStartE2EDuration="7.257620186s" podCreationTimestamp="2025-10-02 08:03:36 +0000 UTC" firstStartedPulling="2025-10-02 08:03:38.152894383 +0000 UTC m=+2839.184840670" lastFinishedPulling="2025-10-02 08:03:42.654539379 +0000 UTC m=+2843.686485666" observedRunningTime="2025-10-02 08:03:43.249119344 +0000 UTC m=+2844.281065631" watchObservedRunningTime="2025-10-02 08:03:43.257620186 +0000 UTC m=+2844.289566473" Oct 02 08:03:43 crc kubenswrapper[4960]: I1002 08:03:43.265668 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z"] Oct 02 08:03:43 crc kubenswrapper[4960]: W1002 08:03:43.266428 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a1cec6e_b3b0_48e7_aa5a_8969b06b069f.slice/crio-b1c71cc39b73f8409b0c4ba743fb970948e1d9b9b4fc21bf9a4df09ac6d3c601 WatchSource:0}: Error finding container b1c71cc39b73f8409b0c4ba743fb970948e1d9b9b4fc21bf9a4df09ac6d3c601: Status 404 returned error can't find the container with id b1c71cc39b73f8409b0c4ba743fb970948e1d9b9b4fc21bf9a4df09ac6d3c601 Oct 02 08:03:44 crc kubenswrapper[4960]: I1002 08:03:44.241911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" event={"ID":"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f","Type":"ContainerStarted","Data":"b1c71cc39b73f8409b0c4ba743fb970948e1d9b9b4fc21bf9a4df09ac6d3c601"} Oct 02 08:03:45 crc kubenswrapper[4960]: I1002 08:03:45.087029 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:45 crc kubenswrapper[4960]: I1002 08:03:45.087406 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:45 crc kubenswrapper[4960]: I1002 08:03:45.146199 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:45 crc kubenswrapper[4960]: I1002 08:03:45.256688 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" event={"ID":"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f","Type":"ContainerStarted","Data":"048c9132b136fecdc4a59dc3dabd0239ca9390cf6ca799fa7a4278094a2dd4c0"} Oct 02 08:03:45 crc kubenswrapper[4960]: I1002 08:03:45.286605 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" podStartSLOduration=2.377130989 podStartE2EDuration="3.286579895s" podCreationTimestamp="2025-10-02 08:03:42 +0000 UTC" firstStartedPulling="2025-10-02 08:03:43.274966757 +0000 UTC m=+2844.306913044" lastFinishedPulling="2025-10-02 08:03:44.184415663 +0000 UTC m=+2845.216361950" observedRunningTime="2025-10-02 08:03:45.282260577 +0000 UTC m=+2846.314206864" watchObservedRunningTime="2025-10-02 08:03:45.286579895 +0000 UTC m=+2846.318526212" Oct 02 08:03:45 crc kubenswrapper[4960]: I1002 08:03:45.321660 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:46 crc kubenswrapper[4960]: I1002 08:03:46.528323 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68bzn"] Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.276723 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-68bzn" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="registry-server" containerID="cri-o://17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00" gracePeriod=2 Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.335460 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.336180 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.393377 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.794872 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.914621 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-catalog-content\") pod \"25873846-5b45-4500-8e66-270f6fdc2722\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.914685 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-utilities\") pod \"25873846-5b45-4500-8e66-270f6fdc2722\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.914973 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xcq8\" (UniqueName: \"kubernetes.io/projected/25873846-5b45-4500-8e66-270f6fdc2722-kube-api-access-7xcq8\") pod \"25873846-5b45-4500-8e66-270f6fdc2722\" (UID: \"25873846-5b45-4500-8e66-270f6fdc2722\") " Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.915548 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-utilities" (OuterVolumeSpecName: "utilities") pod "25873846-5b45-4500-8e66-270f6fdc2722" (UID: "25873846-5b45-4500-8e66-270f6fdc2722"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:03:47 crc kubenswrapper[4960]: I1002 08:03:47.922133 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25873846-5b45-4500-8e66-270f6fdc2722-kube-api-access-7xcq8" (OuterVolumeSpecName: "kube-api-access-7xcq8") pod "25873846-5b45-4500-8e66-270f6fdc2722" (UID: "25873846-5b45-4500-8e66-270f6fdc2722"). InnerVolumeSpecName "kube-api-access-7xcq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.009061 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25873846-5b45-4500-8e66-270f6fdc2722" (UID: "25873846-5b45-4500-8e66-270f6fdc2722"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.016893 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xcq8\" (UniqueName: \"kubernetes.io/projected/25873846-5b45-4500-8e66-270f6fdc2722-kube-api-access-7xcq8\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.016927 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.016939 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25873846-5b45-4500-8e66-270f6fdc2722-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.290269 4960 generic.go:334] "Generic (PLEG): container finished" podID="25873846-5b45-4500-8e66-270f6fdc2722" containerID="17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00" exitCode=0 Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.290403 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68bzn" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.290462 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68bzn" event={"ID":"25873846-5b45-4500-8e66-270f6fdc2722","Type":"ContainerDied","Data":"17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00"} Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.290505 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68bzn" event={"ID":"25873846-5b45-4500-8e66-270f6fdc2722","Type":"ContainerDied","Data":"73e3b8f02ca848f6a1800ac16e27988e5981bd53e456d6641bacbb20f600abba"} Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.290537 4960 scope.go:117] "RemoveContainer" containerID="17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.344326 4960 scope.go:117] "RemoveContainer" containerID="c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.361588 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68bzn"] Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.362170 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-68bzn"] Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.375140 4960 scope.go:117] "RemoveContainer" containerID="809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.384839 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.427636 4960 scope.go:117] "RemoveContainer" containerID="17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00" Oct 02 08:03:48 crc kubenswrapper[4960]: E1002 08:03:48.429300 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00\": container with ID starting with 17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00 not found: ID does not exist" containerID="17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.429369 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00"} err="failed to get container status \"17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00\": rpc error: code = NotFound desc = could not find container \"17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00\": container with ID starting with 17fdadc6055668e3521bcef5772496d8580516f039724aa383d87895474a3d00 not found: ID does not exist" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.429422 4960 scope.go:117] "RemoveContainer" containerID="c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338" Oct 02 08:03:48 crc kubenswrapper[4960]: E1002 08:03:48.431118 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338\": container with ID starting with c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338 not found: ID does not exist" containerID="c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.431223 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338"} err="failed to get container status \"c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338\": rpc error: code = NotFound desc = could not find container \"c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338\": container with ID starting with c65d136af1c5bb6a08f4743cda15e4264ca3afd5c42589b874fcd18f339c1338 not found: ID does not exist" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.431284 4960 scope.go:117] "RemoveContainer" containerID="809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af" Oct 02 08:03:48 crc kubenswrapper[4960]: E1002 08:03:48.431768 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af\": container with ID starting with 809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af not found: ID does not exist" containerID="809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af" Oct 02 08:03:48 crc kubenswrapper[4960]: I1002 08:03:48.432076 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af"} err="failed to get container status \"809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af\": rpc error: code = NotFound desc = could not find container \"809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af\": container with ID starting with 809d6cced6b8eff80146a787706561baf12b720929053f941898b18f74e038af not found: ID does not exist" Oct 02 08:03:49 crc kubenswrapper[4960]: I1002 08:03:49.737928 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zkdjz"] Oct 02 08:03:50 crc kubenswrapper[4960]: I1002 08:03:50.348514 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25873846-5b45-4500-8e66-270f6fdc2722" path="/var/lib/kubelet/pods/25873846-5b45-4500-8e66-270f6fdc2722/volumes" Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.326146 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zkdjz" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="registry-server" containerID="cri-o://174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c" gracePeriod=2 Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.800460 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.919560 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-utilities\") pod \"b58338a1-91dc-4e64-b257-33d487e83d7f\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.919659 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-catalog-content\") pod \"b58338a1-91dc-4e64-b257-33d487e83d7f\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.919698 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj6wg\" (UniqueName: \"kubernetes.io/projected/b58338a1-91dc-4e64-b257-33d487e83d7f-kube-api-access-kj6wg\") pod \"b58338a1-91dc-4e64-b257-33d487e83d7f\" (UID: \"b58338a1-91dc-4e64-b257-33d487e83d7f\") " Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.920580 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-utilities" (OuterVolumeSpecName: "utilities") pod "b58338a1-91dc-4e64-b257-33d487e83d7f" (UID: "b58338a1-91dc-4e64-b257-33d487e83d7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.927107 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b58338a1-91dc-4e64-b257-33d487e83d7f-kube-api-access-kj6wg" (OuterVolumeSpecName: "kube-api-access-kj6wg") pod "b58338a1-91dc-4e64-b257-33d487e83d7f" (UID: "b58338a1-91dc-4e64-b257-33d487e83d7f"). InnerVolumeSpecName "kube-api-access-kj6wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:03:51 crc kubenswrapper[4960]: I1002 08:03:51.968489 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b58338a1-91dc-4e64-b257-33d487e83d7f" (UID: "b58338a1-91dc-4e64-b257-33d487e83d7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.023662 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.023768 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b58338a1-91dc-4e64-b257-33d487e83d7f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.023902 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj6wg\" (UniqueName: \"kubernetes.io/projected/b58338a1-91dc-4e64-b257-33d487e83d7f-kube-api-access-kj6wg\") on node \"crc\" DevicePath \"\"" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.339928 4960 generic.go:334] "Generic (PLEG): container finished" podID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerID="174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c" exitCode=0 Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.340046 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkdjz" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.343087 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkdjz" event={"ID":"b58338a1-91dc-4e64-b257-33d487e83d7f","Type":"ContainerDied","Data":"174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c"} Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.343143 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkdjz" event={"ID":"b58338a1-91dc-4e64-b257-33d487e83d7f","Type":"ContainerDied","Data":"edf3103d12a7b433dee3b08a680af1378eb624ce4e33f59e25a3cb2a51d59eb6"} Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.343168 4960 scope.go:117] "RemoveContainer" containerID="174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.374012 4960 scope.go:117] "RemoveContainer" containerID="2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.382665 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zkdjz"] Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.392430 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zkdjz"] Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.394970 4960 scope.go:117] "RemoveContainer" containerID="6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.458056 4960 scope.go:117] "RemoveContainer" containerID="174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c" Oct 02 08:03:52 crc kubenswrapper[4960]: E1002 08:03:52.458740 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c\": container with ID starting with 174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c not found: ID does not exist" containerID="174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.458829 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c"} err="failed to get container status \"174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c\": rpc error: code = NotFound desc = could not find container \"174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c\": container with ID starting with 174ef99f30f322b5321d8b528706f05974c1fac55ccefad0205a6ec59f42939c not found: ID does not exist" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.458884 4960 scope.go:117] "RemoveContainer" containerID="2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7" Oct 02 08:03:52 crc kubenswrapper[4960]: E1002 08:03:52.459600 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7\": container with ID starting with 2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7 not found: ID does not exist" containerID="2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.459650 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7"} err="failed to get container status \"2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7\": rpc error: code = NotFound desc = could not find container \"2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7\": container with ID starting with 2acdbf232afbc65628a659e94adc424422c58412873c1d7abcb853943cf3b5b7 not found: ID does not exist" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.459683 4960 scope.go:117] "RemoveContainer" containerID="6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627" Oct 02 08:03:52 crc kubenswrapper[4960]: E1002 08:03:52.459968 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627\": container with ID starting with 6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627 not found: ID does not exist" containerID="6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627" Oct 02 08:03:52 crc kubenswrapper[4960]: I1002 08:03:52.460230 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627"} err="failed to get container status \"6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627\": rpc error: code = NotFound desc = could not find container \"6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627\": container with ID starting with 6e3b632d15ed218be6e233cbb7b74ccb9b58ea5ad88082b8fd4ecd09cc738627 not found: ID does not exist" Oct 02 08:03:54 crc kubenswrapper[4960]: I1002 08:03:54.342618 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" path="/var/lib/kubelet/pods/b58338a1-91dc-4e64-b257-33d487e83d7f/volumes" Oct 02 08:03:59 crc kubenswrapper[4960]: I1002 08:03:59.150400 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:03:59 crc kubenswrapper[4960]: I1002 08:03:59.151049 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:04:29 crc kubenswrapper[4960]: I1002 08:04:29.150170 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:04:29 crc kubenswrapper[4960]: I1002 08:04:29.151064 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:04:59 crc kubenswrapper[4960]: I1002 08:04:59.150216 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:04:59 crc kubenswrapper[4960]: I1002 08:04:59.151152 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:04:59 crc kubenswrapper[4960]: I1002 08:04:59.151230 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:04:59 crc kubenswrapper[4960]: I1002 08:04:59.152523 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:04:59 crc kubenswrapper[4960]: I1002 08:04:59.152634 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" gracePeriod=600 Oct 02 08:04:59 crc kubenswrapper[4960]: E1002 08:04:59.300337 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:05:00 crc kubenswrapper[4960]: I1002 08:05:00.097792 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" exitCode=0 Oct 02 08:05:00 crc kubenswrapper[4960]: I1002 08:05:00.097877 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498"} Oct 02 08:05:00 crc kubenswrapper[4960]: I1002 08:05:00.097950 4960 scope.go:117] "RemoveContainer" containerID="697e5417554915e415b18e62ccd2b91fc62b5307191461e4f59ddcb5fd447d6d" Oct 02 08:05:00 crc kubenswrapper[4960]: I1002 08:05:00.099128 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:05:00 crc kubenswrapper[4960]: E1002 08:05:00.099735 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:05:12 crc kubenswrapper[4960]: I1002 08:05:12.329836 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:05:12 crc kubenswrapper[4960]: E1002 08:05:12.330965 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:05:26 crc kubenswrapper[4960]: I1002 08:05:26.330305 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:05:26 crc kubenswrapper[4960]: E1002 08:05:26.331290 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:05:39 crc kubenswrapper[4960]: I1002 08:05:39.331225 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:05:39 crc kubenswrapper[4960]: E1002 08:05:39.332587 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:05:53 crc kubenswrapper[4960]: I1002 08:05:53.330502 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:05:53 crc kubenswrapper[4960]: E1002 08:05:53.333017 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:06:05 crc kubenswrapper[4960]: I1002 08:06:05.330341 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:06:05 crc kubenswrapper[4960]: E1002 08:06:05.333507 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:06:17 crc kubenswrapper[4960]: I1002 08:06:17.331038 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:06:17 crc kubenswrapper[4960]: E1002 08:06:17.336249 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:06:28 crc kubenswrapper[4960]: I1002 08:06:28.330114 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:06:28 crc kubenswrapper[4960]: E1002 08:06:28.331291 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:06:40 crc kubenswrapper[4960]: I1002 08:06:40.338235 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:06:40 crc kubenswrapper[4960]: E1002 08:06:40.339237 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:06:54 crc kubenswrapper[4960]: I1002 08:06:54.330700 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:06:54 crc kubenswrapper[4960]: E1002 08:06:54.332120 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:07:07 crc kubenswrapper[4960]: I1002 08:07:07.329695 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:07:07 crc kubenswrapper[4960]: E1002 08:07:07.330580 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:07:19 crc kubenswrapper[4960]: I1002 08:07:19.330345 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:07:19 crc kubenswrapper[4960]: E1002 08:07:19.331608 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:07:31 crc kubenswrapper[4960]: I1002 08:07:31.330848 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:07:31 crc kubenswrapper[4960]: E1002 08:07:31.332191 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:07:42 crc kubenswrapper[4960]: I1002 08:07:42.330508 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:07:42 crc kubenswrapper[4960]: E1002 08:07:42.332193 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:07:57 crc kubenswrapper[4960]: I1002 08:07:57.330718 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:07:57 crc kubenswrapper[4960]: E1002 08:07:57.331651 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:07:58 crc kubenswrapper[4960]: I1002 08:07:58.268066 4960 generic.go:334] "Generic (PLEG): container finished" podID="4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" containerID="048c9132b136fecdc4a59dc3dabd0239ca9390cf6ca799fa7a4278094a2dd4c0" exitCode=0 Oct 02 08:07:58 crc kubenswrapper[4960]: I1002 08:07:58.268211 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" event={"ID":"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f","Type":"ContainerDied","Data":"048c9132b136fecdc4a59dc3dabd0239ca9390cf6ca799fa7a4278094a2dd4c0"} Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.743762 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.850299 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.850361 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ssh-key\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851182 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-0\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851224 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-custom-ceph-combined-ca-bundle\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851358 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-1\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851384 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-0\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851448 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-extra-config-0\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851523 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-inventory\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851555 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmr6t\" (UniqueName: \"kubernetes.io/projected/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-kube-api-access-qmr6t\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851578 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph-nova-0\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.851603 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-1\") pod \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\" (UID: \"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f\") " Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.860033 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph" (OuterVolumeSpecName: "ceph") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.861159 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-kube-api-access-qmr6t" (OuterVolumeSpecName: "kube-api-access-qmr6t") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "kube-api-access-qmr6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.865069 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.888050 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.889167 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.895289 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.896725 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.897535 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.898039 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.902578 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.933657 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-inventory" (OuterVolumeSpecName: "inventory") pod "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" (UID: "4a1cec6e-b3b0-48e7-aa5a-8969b06b069f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955206 4960 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955258 4960 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955272 4960 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955284 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955298 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmr6t\" (UniqueName: \"kubernetes.io/projected/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-kube-api-access-qmr6t\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955311 4960 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955324 4960 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955338 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955351 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955364 4960 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:07:59 crc kubenswrapper[4960]: I1002 08:07:59.955373 4960 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a1cec6e-b3b0-48e7-aa5a-8969b06b069f-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:00 crc kubenswrapper[4960]: I1002 08:08:00.305811 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" event={"ID":"4a1cec6e-b3b0-48e7-aa5a-8969b06b069f","Type":"ContainerDied","Data":"b1c71cc39b73f8409b0c4ba743fb970948e1d9b9b4fc21bf9a4df09ac6d3c601"} Oct 02 08:08:00 crc kubenswrapper[4960]: I1002 08:08:00.305868 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1c71cc39b73f8409b0c4ba743fb970948e1d9b9b4fc21bf9a4df09ac6d3c601" Oct 02 08:08:00 crc kubenswrapper[4960]: I1002 08:08:00.305950 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z" Oct 02 08:08:08 crc kubenswrapper[4960]: I1002 08:08:08.332262 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:08:08 crc kubenswrapper[4960]: E1002 08:08:08.333660 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.797726 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 02 08:08:14 crc kubenswrapper[4960]: E1002 08:08:14.798906 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="extract-content" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.798926 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="extract-content" Oct 02 08:08:14 crc kubenswrapper[4960]: E1002 08:08:14.798949 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.798958 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 02 08:08:14 crc kubenswrapper[4960]: E1002 08:08:14.799013 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="extract-utilities" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799022 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="extract-utilities" Oct 02 08:08:14 crc kubenswrapper[4960]: E1002 08:08:14.799041 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="extract-content" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799048 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="extract-content" Oct 02 08:08:14 crc kubenswrapper[4960]: E1002 08:08:14.799065 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="registry-server" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799072 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="registry-server" Oct 02 08:08:14 crc kubenswrapper[4960]: E1002 08:08:14.799083 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="registry-server" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799090 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="registry-server" Oct 02 08:08:14 crc kubenswrapper[4960]: E1002 08:08:14.799106 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="extract-utilities" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799113 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="extract-utilities" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799318 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a1cec6e-b3b0-48e7-aa5a-8969b06b069f" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799335 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="25873846-5b45-4500-8e66-270f6fdc2722" containerName="registry-server" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.799357 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b58338a1-91dc-4e64-b257-33d487e83d7f" containerName="registry-server" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.800552 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.803161 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.803314 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.818546 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.820579 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.824252 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.892515 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920355 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920448 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-config-data\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920507 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-sys\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920606 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqvmj\" (UniqueName: \"kubernetes.io/projected/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-kube-api-access-qqvmj\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920717 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920747 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920775 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-lib-modules\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zjvw\" (UniqueName: \"kubernetes.io/projected/0a9c86d8-b033-44fa-9991-08a9aa629b32-kube-api-access-9zjvw\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920858 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920905 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920921 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920944 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.920987 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921023 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921053 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921076 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921096 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921120 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921149 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-run\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921178 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-dev\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921282 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921306 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921331 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921350 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921399 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-scripts\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921427 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0a9c86d8-b033-44fa-9991-08a9aa629b32-ceph\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921470 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-run\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921506 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921533 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921591 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.921635 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:14 crc kubenswrapper[4960]: I1002 08:08:14.936752 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.023892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.023948 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.023988 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-lib-modules\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024023 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zjvw\" (UniqueName: \"kubernetes.io/projected/0a9c86d8-b033-44fa-9991-08a9aa629b32-kube-api-access-9zjvw\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024040 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024065 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024081 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024105 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024124 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024147 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024168 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025319 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025344 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025361 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024244 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-lib-modules\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024330 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025415 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025465 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024451 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-run\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025597 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-dev\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025408 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-run\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024411 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024444 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024444 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.024437 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025429 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025707 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-dev\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025825 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025846 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025885 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025911 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-scripts\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025929 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0a9c86d8-b033-44fa-9991-08a9aa629b32-ceph\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.025953 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-run\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026017 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026035 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026071 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026096 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026120 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-config-data\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026169 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-sys\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026216 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqvmj\" (UniqueName: \"kubernetes.io/projected/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-kube-api-access-qqvmj\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026236 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026318 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026342 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.026363 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.028106 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-sys\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.028135 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a9c86d8-b033-44fa-9991-08a9aa629b32-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.028873 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.029008 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-run\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.033598 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.034193 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.035021 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.041743 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.042306 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-config-data\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.043623 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.043642 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.045588 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zjvw\" (UniqueName: \"kubernetes.io/projected/0a9c86d8-b033-44fa-9991-08a9aa629b32-kube-api-access-9zjvw\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.047105 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0a9c86d8-b033-44fa-9991-08a9aa629b32-ceph\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.047725 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqvmj\" (UniqueName: \"kubernetes.io/projected/3bebb599-2d7f-4d91-8cf1-eda510fd14f6-kube-api-access-qqvmj\") pod \"cinder-volume-volume1-0\" (UID: \"3bebb599-2d7f-4d91-8cf1-eda510fd14f6\") " pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.048142 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-scripts\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.048346 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a9c86d8-b033-44fa-9991-08a9aa629b32-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0a9c86d8-b033-44fa-9991-08a9aa629b32\") " pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.126085 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.215449 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.424310 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-7h55n"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.426394 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-7h55n" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.439525 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-7h55n"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.537614 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqvlq\" (UniqueName: \"kubernetes.io/projected/0f90db36-1706-4086-8289-91cdf46731ac-kube-api-access-rqvlq\") pod \"manila-db-create-7h55n\" (UID: \"0f90db36-1706-4086-8289-91cdf46731ac\") " pod="openstack/manila-db-create-7h55n" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.639651 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.641059 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqvlq\" (UniqueName: \"kubernetes.io/projected/0f90db36-1706-4086-8289-91cdf46731ac-kube-api-access-rqvlq\") pod \"manila-db-create-7h55n\" (UID: \"0f90db36-1706-4086-8289-91cdf46731ac\") " pod="openstack/manila-db-create-7h55n" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.641368 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.645628 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.645901 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4gqq6" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.646101 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.647968 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.654047 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.686102 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqvlq\" (UniqueName: \"kubernetes.io/projected/0f90db36-1706-4086-8289-91cdf46731ac-kube-api-access-rqvlq\") pod \"manila-db-create-7h55n\" (UID: \"0f90db36-1706-4086-8289-91cdf46731ac\") " pod="openstack/manila-db-create-7h55n" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.716990 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.719268 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.723284 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.723627 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.724691 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742604 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-scripts\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742676 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-config-data\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742772 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8da35d5f-3f76-4e03-9df7-7d26b23c601d-logs\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742833 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742869 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8da35d5f-3f76-4e03-9df7-7d26b23c601d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742897 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742943 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8da35d5f-3f76-4e03-9df7-7d26b23c601d-ceph\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.742974 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.743001 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-956hh\" (UniqueName: \"kubernetes.io/projected/8da35d5f-3f76-4e03-9df7-7d26b23c601d-kube-api-access-956hh\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.749130 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-7h55n" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.787576 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.803710 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845269 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845348 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8da35d5f-3f76-4e03-9df7-7d26b23c601d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845388 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845420 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845462 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnjvx\" (UniqueName: \"kubernetes.io/projected/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-kube-api-access-xnjvx\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845501 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8da35d5f-3f76-4e03-9df7-7d26b23c601d-ceph\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845545 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845606 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-956hh\" (UniqueName: \"kubernetes.io/projected/8da35d5f-3f76-4e03-9df7-7d26b23c601d-kube-api-access-956hh\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845649 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845686 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-scripts\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845724 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-config-data\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845753 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845817 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8da35d5f-3f76-4e03-9df7-7d26b23c601d-logs\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845855 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845956 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.845986 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.846895 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.846910 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8da35d5f-3f76-4e03-9df7-7d26b23c601d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.848764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8da35d5f-3f76-4e03-9df7-7d26b23c601d-logs\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.852127 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8da35d5f-3f76-4e03-9df7-7d26b23c601d-ceph\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.855371 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.855416 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-config-data\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.857817 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.858295 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da35d5f-3f76-4e03-9df7-7d26b23c601d-scripts\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.876560 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-956hh\" (UniqueName: \"kubernetes.io/projected/8da35d5f-3f76-4e03-9df7-7d26b23c601d-kube-api-access-956hh\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.897656 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8da35d5f-3f76-4e03-9df7-7d26b23c601d\") " pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.901448 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 02 08:08:15 crc kubenswrapper[4960]: W1002 08:08:15.905658 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a9c86d8_b033_44fa_9991_08a9aa629b32.slice/crio-3a9e148c505d32a7124b9727a4f527d0785f1f1474e0f797f45c97187be9f808 WatchSource:0}: Error finding container 3a9e148c505d32a7124b9727a4f527d0785f1f1474e0f797f45c97187be9f808: Status 404 returned error can't find the container with id 3a9e148c505d32a7124b9727a4f527d0785f1f1474e0f797f45c97187be9f808 Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.948827 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.948891 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnjvx\" (UniqueName: \"kubernetes.io/projected/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-kube-api-access-xnjvx\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.948962 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.949016 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.949299 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.949347 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.949391 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.949422 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.949461 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.950357 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.951535 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.952097 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.967723 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.970023 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.974434 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.977123 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.979190 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.970861 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:15 crc kubenswrapper[4960]: I1002 08:08:15.993882 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnjvx\" (UniqueName: \"kubernetes.io/projected/7bde8f5a-7f1c-4c7d-a52c-ee1269736978-kube-api-access-xnjvx\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:16 crc kubenswrapper[4960]: I1002 08:08:16.018664 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bde8f5a-7f1c-4c7d-a52c-ee1269736978\") " pod="openstack/glance-default-internal-api-0" Oct 02 08:08:16 crc kubenswrapper[4960]: I1002 08:08:16.056495 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:16 crc kubenswrapper[4960]: I1002 08:08:16.268445 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-7h55n"] Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:16.510268 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3bebb599-2d7f-4d91-8cf1-eda510fd14f6","Type":"ContainerStarted","Data":"5d0f131f3ebf63a17e2decffcd2f7208455a1fce31a1d287bfc851a85f863f8a"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:16.511789 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-7h55n" event={"ID":"0f90db36-1706-4086-8289-91cdf46731ac","Type":"ContainerStarted","Data":"9d6baa17c3e3c7f230106d3b15cc51ff8d19822a70c3344e5297d2d359ce85bd"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:16.518116 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0a9c86d8-b033-44fa-9991-08a9aa629b32","Type":"ContainerStarted","Data":"3a9e148c505d32a7124b9727a4f527d0785f1f1474e0f797f45c97187be9f808"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:16.848139 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:16.941524 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:17.534041 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8da35d5f-3f76-4e03-9df7-7d26b23c601d","Type":"ContainerStarted","Data":"e66e4ef200b2d159d4af9db377273336fc80a696f4017347770fed4016d85d39"} Oct 02 08:08:20 crc kubenswrapper[4960]: W1002 08:08:19.107805 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bde8f5a_7f1c_4c7d_a52c_ee1269736978.slice/crio-f2a0e41c56b08c790dbe51bb4b06573db224f4db1582daaf1d582f8586c149bf WatchSource:0}: Error finding container f2a0e41c56b08c790dbe51bb4b06573db224f4db1582daaf1d582f8586c149bf: Status 404 returned error can't find the container with id f2a0e41c56b08c790dbe51bb4b06573db224f4db1582daaf1d582f8586c149bf Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:19.561257 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bde8f5a-7f1c-4c7d-a52c-ee1269736978","Type":"ContainerStarted","Data":"f2a0e41c56b08c790dbe51bb4b06573db224f4db1582daaf1d582f8586c149bf"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:19.563825 4960 generic.go:334] "Generic (PLEG): container finished" podID="0f90db36-1706-4086-8289-91cdf46731ac" containerID="f07f2966e81ba9ed92661bf3a39a7bd1d865abae4a9a0f05916d398658c8a6ac" exitCode=0 Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:19.563874 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-7h55n" event={"ID":"0f90db36-1706-4086-8289-91cdf46731ac","Type":"ContainerDied","Data":"f07f2966e81ba9ed92661bf3a39a7bd1d865abae4a9a0f05916d398658c8a6ac"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:20.350780 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:08:20 crc kubenswrapper[4960]: E1002 08:08:20.352166 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:20.584946 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3bebb599-2d7f-4d91-8cf1-eda510fd14f6","Type":"ContainerStarted","Data":"817e34d11397a0cb33db843608cfceea085845299fdaa97512b13feb4d2c6c11"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:20.587845 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0a9c86d8-b033-44fa-9991-08a9aa629b32","Type":"ContainerStarted","Data":"6179b50edbbd7971ff2a3bc83121f0e46ff81131660d9321f785177ac7877770"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:20.592595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bde8f5a-7f1c-4c7d-a52c-ee1269736978","Type":"ContainerStarted","Data":"20f4bd51cdbad4c45654013916b29a54e705f8954cae89f3e10d1d27bab67a4d"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:20.597268 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8da35d5f-3f76-4e03-9df7-7d26b23c601d","Type":"ContainerStarted","Data":"e9bfd5ef3eee5709ffa2c667c96a9318ce4de67113b882b8448b7f0d5a2c53aa"} Oct 02 08:08:20 crc kubenswrapper[4960]: I1002 08:08:20.900737 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-7h55n" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.025773 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqvlq\" (UniqueName: \"kubernetes.io/projected/0f90db36-1706-4086-8289-91cdf46731ac-kube-api-access-rqvlq\") pod \"0f90db36-1706-4086-8289-91cdf46731ac\" (UID: \"0f90db36-1706-4086-8289-91cdf46731ac\") " Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.041229 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f90db36-1706-4086-8289-91cdf46731ac-kube-api-access-rqvlq" (OuterVolumeSpecName: "kube-api-access-rqvlq") pod "0f90db36-1706-4086-8289-91cdf46731ac" (UID: "0f90db36-1706-4086-8289-91cdf46731ac"). InnerVolumeSpecName "kube-api-access-rqvlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.128049 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqvlq\" (UniqueName: \"kubernetes.io/projected/0f90db36-1706-4086-8289-91cdf46731ac-kube-api-access-rqvlq\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.613601 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bde8f5a-7f1c-4c7d-a52c-ee1269736978","Type":"ContainerStarted","Data":"0c310de1097518be3255f438d2a6faffc262a908481b9e87c4855f3b30167756"} Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.616079 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8da35d5f-3f76-4e03-9df7-7d26b23c601d","Type":"ContainerStarted","Data":"5e424b8e002f9d6c838105eb82cd9562da3b610234cd3fa69b0ac41e112fc9e2"} Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.619436 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3bebb599-2d7f-4d91-8cf1-eda510fd14f6","Type":"ContainerStarted","Data":"693605c0fad9591d5d422295e56af456d52f3bb2039bac929baa246259e0248a"} Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.625972 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-7h55n" event={"ID":"0f90db36-1706-4086-8289-91cdf46731ac","Type":"ContainerDied","Data":"9d6baa17c3e3c7f230106d3b15cc51ff8d19822a70c3344e5297d2d359ce85bd"} Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.626093 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d6baa17c3e3c7f230106d3b15cc51ff8d19822a70c3344e5297d2d359ce85bd" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.626028 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-7h55n" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.629108 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0a9c86d8-b033-44fa-9991-08a9aa629b32","Type":"ContainerStarted","Data":"8e4fe2bbe717493f252e2012be10f8d41aacac7f49ccefb3801565b701da0201"} Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.669390 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.669366194 podStartE2EDuration="7.669366194s" podCreationTimestamp="2025-10-02 08:08:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:08:21.651593299 +0000 UTC m=+3122.683539586" watchObservedRunningTime="2025-10-02 08:08:21.669366194 +0000 UTC m=+3122.701312481" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.694608 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.71674574 podStartE2EDuration="7.694586039s" podCreationTimestamp="2025-10-02 08:08:14 +0000 UTC" firstStartedPulling="2025-10-02 08:08:15.909022397 +0000 UTC m=+3116.940968684" lastFinishedPulling="2025-10-02 08:08:19.886862676 +0000 UTC m=+3120.918808983" observedRunningTime="2025-10-02 08:08:21.690847829 +0000 UTC m=+3122.722794116" watchObservedRunningTime="2025-10-02 08:08:21.694586039 +0000 UTC m=+3122.726532336" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.728133 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.728108046 podStartE2EDuration="7.728108046s" podCreationTimestamp="2025-10-02 08:08:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:08:21.720429881 +0000 UTC m=+3122.752376208" watchObservedRunningTime="2025-10-02 08:08:21.728108046 +0000 UTC m=+3122.760054323" Oct 02 08:08:21 crc kubenswrapper[4960]: I1002 08:08:21.757417 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.6722367399999998 podStartE2EDuration="7.757389189s" podCreationTimestamp="2025-10-02 08:08:14 +0000 UTC" firstStartedPulling="2025-10-02 08:08:15.803501025 +0000 UTC m=+3116.835447312" lastFinishedPulling="2025-10-02 08:08:19.888653474 +0000 UTC m=+3120.920599761" observedRunningTime="2025-10-02 08:08:21.750867124 +0000 UTC m=+3122.782813411" watchObservedRunningTime="2025-10-02 08:08:21.757389189 +0000 UTC m=+3122.789335486" Oct 02 08:08:25 crc kubenswrapper[4960]: I1002 08:08:25.127517 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:25 crc kubenswrapper[4960]: I1002 08:08:25.216962 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 02 08:08:25 crc kubenswrapper[4960]: I1002 08:08:25.409447 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 02 08:08:25 crc kubenswrapper[4960]: I1002 08:08:25.479398 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 02 08:08:25 crc kubenswrapper[4960]: I1002 08:08:25.980565 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 08:08:25 crc kubenswrapper[4960]: I1002 08:08:25.980640 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.035839 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.048430 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.057805 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.059510 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.119530 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.131771 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.713864 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.717167 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.717310 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 08:08:26 crc kubenswrapper[4960]: I1002 08:08:26.717329 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 08:08:28 crc kubenswrapper[4960]: I1002 08:08:28.730843 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 08:08:28 crc kubenswrapper[4960]: I1002 08:08:28.731399 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 08:08:28 crc kubenswrapper[4960]: I1002 08:08:28.983749 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.043094 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.043240 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.060820 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.072370 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.294922 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-8be1-account-create-gpbfg"] Oct 02 08:08:29 crc kubenswrapper[4960]: E1002 08:08:29.295461 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f90db36-1706-4086-8289-91cdf46731ac" containerName="mariadb-database-create" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.295481 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f90db36-1706-4086-8289-91cdf46731ac" containerName="mariadb-database-create" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.295723 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f90db36-1706-4086-8289-91cdf46731ac" containerName="mariadb-database-create" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.296573 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8be1-account-create-gpbfg" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.301795 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.310420 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-8be1-account-create-gpbfg"] Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.440203 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2qxz\" (UniqueName: \"kubernetes.io/projected/3fc4a490-fbec-448d-bbcf-e436060d9497-kube-api-access-c2qxz\") pod \"manila-8be1-account-create-gpbfg\" (UID: \"3fc4a490-fbec-448d-bbcf-e436060d9497\") " pod="openstack/manila-8be1-account-create-gpbfg" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.543358 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2qxz\" (UniqueName: \"kubernetes.io/projected/3fc4a490-fbec-448d-bbcf-e436060d9497-kube-api-access-c2qxz\") pod \"manila-8be1-account-create-gpbfg\" (UID: \"3fc4a490-fbec-448d-bbcf-e436060d9497\") " pod="openstack/manila-8be1-account-create-gpbfg" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.573587 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2qxz\" (UniqueName: \"kubernetes.io/projected/3fc4a490-fbec-448d-bbcf-e436060d9497-kube-api-access-c2qxz\") pod \"manila-8be1-account-create-gpbfg\" (UID: \"3fc4a490-fbec-448d-bbcf-e436060d9497\") " pod="openstack/manila-8be1-account-create-gpbfg" Oct 02 08:08:29 crc kubenswrapper[4960]: I1002 08:08:29.631778 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8be1-account-create-gpbfg" Oct 02 08:08:30 crc kubenswrapper[4960]: I1002 08:08:30.178039 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-8be1-account-create-gpbfg"] Oct 02 08:08:30 crc kubenswrapper[4960]: I1002 08:08:30.761605 4960 generic.go:334] "Generic (PLEG): container finished" podID="3fc4a490-fbec-448d-bbcf-e436060d9497" containerID="204a8ce30c11934c4044fe935f2d320cf579434ab72b53513c4abb86de0ea028" exitCode=0 Oct 02 08:08:30 crc kubenswrapper[4960]: I1002 08:08:30.761798 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-8be1-account-create-gpbfg" event={"ID":"3fc4a490-fbec-448d-bbcf-e436060d9497","Type":"ContainerDied","Data":"204a8ce30c11934c4044fe935f2d320cf579434ab72b53513c4abb86de0ea028"} Oct 02 08:08:30 crc kubenswrapper[4960]: I1002 08:08:30.762440 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-8be1-account-create-gpbfg" event={"ID":"3fc4a490-fbec-448d-bbcf-e436060d9497","Type":"ContainerStarted","Data":"a0ff242f6255f82f37d49e4462dbeea8dcf1207365d0c77dd25797ff9df0b861"} Oct 02 08:08:32 crc kubenswrapper[4960]: I1002 08:08:32.134699 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8be1-account-create-gpbfg" Oct 02 08:08:32 crc kubenswrapper[4960]: I1002 08:08:32.327284 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2qxz\" (UniqueName: \"kubernetes.io/projected/3fc4a490-fbec-448d-bbcf-e436060d9497-kube-api-access-c2qxz\") pod \"3fc4a490-fbec-448d-bbcf-e436060d9497\" (UID: \"3fc4a490-fbec-448d-bbcf-e436060d9497\") " Oct 02 08:08:32 crc kubenswrapper[4960]: I1002 08:08:32.341146 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc4a490-fbec-448d-bbcf-e436060d9497-kube-api-access-c2qxz" (OuterVolumeSpecName: "kube-api-access-c2qxz") pod "3fc4a490-fbec-448d-bbcf-e436060d9497" (UID: "3fc4a490-fbec-448d-bbcf-e436060d9497"). InnerVolumeSpecName "kube-api-access-c2qxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:08:32 crc kubenswrapper[4960]: I1002 08:08:32.432295 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2qxz\" (UniqueName: \"kubernetes.io/projected/3fc4a490-fbec-448d-bbcf-e436060d9497-kube-api-access-c2qxz\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:32 crc kubenswrapper[4960]: I1002 08:08:32.791056 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-8be1-account-create-gpbfg" event={"ID":"3fc4a490-fbec-448d-bbcf-e436060d9497","Type":"ContainerDied","Data":"a0ff242f6255f82f37d49e4462dbeea8dcf1207365d0c77dd25797ff9df0b861"} Oct 02 08:08:32 crc kubenswrapper[4960]: I1002 08:08:32.791126 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0ff242f6255f82f37d49e4462dbeea8dcf1207365d0c77dd25797ff9df0b861" Oct 02 08:08:32 crc kubenswrapper[4960]: I1002 08:08:32.791219 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-8be1-account-create-gpbfg" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.330317 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:08:34 crc kubenswrapper[4960]: E1002 08:08:34.331242 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.672883 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-czszs"] Oct 02 08:08:34 crc kubenswrapper[4960]: E1002 08:08:34.673877 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc4a490-fbec-448d-bbcf-e436060d9497" containerName="mariadb-account-create" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.673904 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc4a490-fbec-448d-bbcf-e436060d9497" containerName="mariadb-account-create" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.674183 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc4a490-fbec-448d-bbcf-e436060d9497" containerName="mariadb-account-create" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.675135 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.687889 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-nv655" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.707507 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.755360 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-czszs"] Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.808486 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-config-data\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.809117 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-job-config-data\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.809459 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-combined-ca-bundle\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.809789 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf4fq\" (UniqueName: \"kubernetes.io/projected/72581d48-c6bf-4140-a8d1-11c058d33638-kube-api-access-vf4fq\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.911905 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf4fq\" (UniqueName: \"kubernetes.io/projected/72581d48-c6bf-4140-a8d1-11c058d33638-kube-api-access-vf4fq\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.912010 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-config-data\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.912087 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-job-config-data\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.912139 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-combined-ca-bundle\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.918857 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-combined-ca-bundle\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.920959 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-config-data\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.934656 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-job-config-data\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:34 crc kubenswrapper[4960]: I1002 08:08:34.944747 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf4fq\" (UniqueName: \"kubernetes.io/projected/72581d48-c6bf-4140-a8d1-11c058d33638-kube-api-access-vf4fq\") pod \"manila-db-sync-czszs\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " pod="openstack/manila-db-sync-czszs" Oct 02 08:08:35 crc kubenswrapper[4960]: I1002 08:08:35.019868 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-czszs" Oct 02 08:08:35 crc kubenswrapper[4960]: I1002 08:08:35.750453 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-czszs"] Oct 02 08:08:35 crc kubenswrapper[4960]: I1002 08:08:35.827011 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-czszs" event={"ID":"72581d48-c6bf-4140-a8d1-11c058d33638","Type":"ContainerStarted","Data":"dd75bee3351a6d92197654ec541a806d89da905a1a33c0f91a2ce68c12f7b563"} Oct 02 08:08:40 crc kubenswrapper[4960]: I1002 08:08:40.903107 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-czszs" event={"ID":"72581d48-c6bf-4140-a8d1-11c058d33638","Type":"ContainerStarted","Data":"b9dff3d366277357551605ff08101eb77268824acc29c1945e0284d2d04a367d"} Oct 02 08:08:40 crc kubenswrapper[4960]: I1002 08:08:40.929417 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-czszs" podStartSLOduration=2.889264567 podStartE2EDuration="6.929398971s" podCreationTimestamp="2025-10-02 08:08:34 +0000 UTC" firstStartedPulling="2025-10-02 08:08:35.757287358 +0000 UTC m=+3136.789233655" lastFinishedPulling="2025-10-02 08:08:39.797421772 +0000 UTC m=+3140.829368059" observedRunningTime="2025-10-02 08:08:40.924531171 +0000 UTC m=+3141.956477458" watchObservedRunningTime="2025-10-02 08:08:40.929398971 +0000 UTC m=+3141.961345258" Oct 02 08:08:47 crc kubenswrapper[4960]: I1002 08:08:47.331341 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:08:47 crc kubenswrapper[4960]: E1002 08:08:47.332560 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:08:51 crc kubenswrapper[4960]: I1002 08:08:51.004437 4960 generic.go:334] "Generic (PLEG): container finished" podID="72581d48-c6bf-4140-a8d1-11c058d33638" containerID="b9dff3d366277357551605ff08101eb77268824acc29c1945e0284d2d04a367d" exitCode=0 Oct 02 08:08:51 crc kubenswrapper[4960]: I1002 08:08:51.004552 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-czszs" event={"ID":"72581d48-c6bf-4140-a8d1-11c058d33638","Type":"ContainerDied","Data":"b9dff3d366277357551605ff08101eb77268824acc29c1945e0284d2d04a367d"} Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.457569 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-czszs" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.562737 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-config-data\") pod \"72581d48-c6bf-4140-a8d1-11c058d33638\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.562933 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf4fq\" (UniqueName: \"kubernetes.io/projected/72581d48-c6bf-4140-a8d1-11c058d33638-kube-api-access-vf4fq\") pod \"72581d48-c6bf-4140-a8d1-11c058d33638\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.563040 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-job-config-data\") pod \"72581d48-c6bf-4140-a8d1-11c058d33638\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.563126 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-combined-ca-bundle\") pod \"72581d48-c6bf-4140-a8d1-11c058d33638\" (UID: \"72581d48-c6bf-4140-a8d1-11c058d33638\") " Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.571602 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72581d48-c6bf-4140-a8d1-11c058d33638-kube-api-access-vf4fq" (OuterVolumeSpecName: "kube-api-access-vf4fq") pod "72581d48-c6bf-4140-a8d1-11c058d33638" (UID: "72581d48-c6bf-4140-a8d1-11c058d33638"). InnerVolumeSpecName "kube-api-access-vf4fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.576964 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "72581d48-c6bf-4140-a8d1-11c058d33638" (UID: "72581d48-c6bf-4140-a8d1-11c058d33638"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.579895 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-config-data" (OuterVolumeSpecName: "config-data") pod "72581d48-c6bf-4140-a8d1-11c058d33638" (UID: "72581d48-c6bf-4140-a8d1-11c058d33638"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.596645 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72581d48-c6bf-4140-a8d1-11c058d33638" (UID: "72581d48-c6bf-4140-a8d1-11c058d33638"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.665636 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf4fq\" (UniqueName: \"kubernetes.io/projected/72581d48-c6bf-4140-a8d1-11c058d33638-kube-api-access-vf4fq\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.665684 4960 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.665695 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:52 crc kubenswrapper[4960]: I1002 08:08:52.665708 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72581d48-c6bf-4140-a8d1-11c058d33638-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.026715 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-czszs" event={"ID":"72581d48-c6bf-4140-a8d1-11c058d33638","Type":"ContainerDied","Data":"dd75bee3351a6d92197654ec541a806d89da905a1a33c0f91a2ce68c12f7b563"} Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.027219 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd75bee3351a6d92197654ec541a806d89da905a1a33c0f91a2ce68c12f7b563" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.026783 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-czszs" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.367671 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:08:53 crc kubenswrapper[4960]: E1002 08:08:53.368221 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72581d48-c6bf-4140-a8d1-11c058d33638" containerName="manila-db-sync" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.368239 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="72581d48-c6bf-4140-a8d1-11c058d33638" containerName="manila-db-sync" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.368413 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="72581d48-c6bf-4140-a8d1-11c058d33638" containerName="manila-db-sync" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.377703 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.382357 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.382601 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.382644 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.382824 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-nv655" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.400462 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.475094 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.476692 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.479158 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.482985 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2695413-bac9-4098-8399-72819bda9615-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.483076 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.483132 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.483151 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.483201 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r57wf\" (UniqueName: \"kubernetes.io/projected/d2695413-bac9-4098-8399-72819bda9615-kube-api-access-r57wf\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.483259 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-scripts\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.499871 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.571082 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6785d65565-vfxf8"] Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.573517 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.585800 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-scripts\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.585850 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-ceph\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.585873 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.585911 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2695413-bac9-4098-8399-72819bda9615-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.585943 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.585965 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586011 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586052 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586078 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586115 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586133 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586165 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dsql\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-kube-api-access-6dsql\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586190 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-scripts\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.586217 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r57wf\" (UniqueName: \"kubernetes.io/projected/d2695413-bac9-4098-8399-72819bda9615-kube-api-access-r57wf\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.587739 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2695413-bac9-4098-8399-72819bda9615-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.592170 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.596912 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.598424 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-scripts\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.614048 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6785d65565-vfxf8"] Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.618267 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.644126 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r57wf\" (UniqueName: \"kubernetes.io/projected/d2695413-bac9-4098-8399-72819bda9615-kube-api-access-r57wf\") pod \"manila-scheduler-0\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691112 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-scripts\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbtzq\" (UniqueName: \"kubernetes.io/projected/b169e295-17c4-457d-8832-bb79f85eb5cb-kube-api-access-jbtzq\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691259 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-ceph\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691275 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691324 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691343 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691360 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-openstack-edpm-ipam\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691386 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-dns-svc\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691402 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691417 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691454 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-config\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691496 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691543 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.691584 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dsql\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-kube-api-access-6dsql\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.693060 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.695283 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-scripts\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.695353 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.697136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.697459 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-ceph\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.700894 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.701053 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.703128 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.710601 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dsql\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-kube-api-access-6dsql\") pod \"manila-share-share1-0\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.776426 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.778082 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.785574 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.787452 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.794438 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbtzq\" (UniqueName: \"kubernetes.io/projected/b169e295-17c4-457d-8832-bb79f85eb5cb-kube-api-access-jbtzq\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.794538 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-openstack-edpm-ipam\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.794573 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.794594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-dns-svc\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.794639 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-config\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.794709 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.795820 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-openstack-edpm-ipam\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.796609 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-dns-svc\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.796638 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.796870 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-config\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.797183 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.797218 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b169e295-17c4-457d-8832-bb79f85eb5cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.833256 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbtzq\" (UniqueName: \"kubernetes.io/projected/b169e295-17c4-457d-8832-bb79f85eb5cb-kube-api-access-jbtzq\") pod \"dnsmasq-dns-6785d65565-vfxf8\" (UID: \"b169e295-17c4-457d-8832-bb79f85eb5cb\") " pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.885243 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.896525 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data-custom\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.896572 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.896612 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0be16689-f82d-4473-898b-a76cb6885219-etc-machine-id\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.896639 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.896654 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be16689-f82d-4473-898b-a76cb6885219-logs\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.896702 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-scripts\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.896760 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctb6c\" (UniqueName: \"kubernetes.io/projected/0be16689-f82d-4473-898b-a76cb6885219-kube-api-access-ctb6c\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.998307 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-scripts\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.998429 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctb6c\" (UniqueName: \"kubernetes.io/projected/0be16689-f82d-4473-898b-a76cb6885219-kube-api-access-ctb6c\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.998574 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data-custom\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.998605 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.998651 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0be16689-f82d-4473-898b-a76cb6885219-etc-machine-id\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.998676 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:53 crc kubenswrapper[4960]: I1002 08:08:53.998690 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be16689-f82d-4473-898b-a76cb6885219-logs\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:53.999234 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be16689-f82d-4473-898b-a76cb6885219-logs\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.000198 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0be16689-f82d-4473-898b-a76cb6885219-etc-machine-id\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.004366 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.006635 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data-custom\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.006819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.013697 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-scripts\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.025782 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctb6c\" (UniqueName: \"kubernetes.io/projected/0be16689-f82d-4473-898b-a76cb6885219-kube-api-access-ctb6c\") pod \"manila-api-0\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.215595 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.309779 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.792150 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6785d65565-vfxf8"] Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.836493 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:08:54 crc kubenswrapper[4960]: I1002 08:08:54.913338 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:55 crc kubenswrapper[4960]: I1002 08:08:55.068092 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0be16689-f82d-4473-898b-a76cb6885219","Type":"ContainerStarted","Data":"d7d66c51ea1947bdc5d6dd505a450da650d13b923a27b9e97d6f1acfad157a4e"} Oct 02 08:08:55 crc kubenswrapper[4960]: I1002 08:08:55.069223 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"d2695413-bac9-4098-8399-72819bda9615","Type":"ContainerStarted","Data":"869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8"} Oct 02 08:08:55 crc kubenswrapper[4960]: I1002 08:08:55.070321 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ad1f362d-1d99-4948-a74e-cb656fbc3825","Type":"ContainerStarted","Data":"0fa97d936f32caf1b20877374200e634589d2e3e797867125a24c7ec8c502fd2"} Oct 02 08:08:55 crc kubenswrapper[4960]: I1002 08:08:55.072016 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" event={"ID":"b169e295-17c4-457d-8832-bb79f85eb5cb","Type":"ContainerStarted","Data":"3f1af90b81819610862396d4334354d6ce8bfbcecfeaf785d0ff7265c02f8fc9"} Oct 02 08:08:56 crc kubenswrapper[4960]: I1002 08:08:56.089563 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"d2695413-bac9-4098-8399-72819bda9615","Type":"ContainerStarted","Data":"083cc1cd63078041bcb54566f3f1f0878218809774cf45d62c9d1792520b4aa3"} Oct 02 08:08:56 crc kubenswrapper[4960]: I1002 08:08:56.099486 4960 generic.go:334] "Generic (PLEG): container finished" podID="b169e295-17c4-457d-8832-bb79f85eb5cb" containerID="6c6f587982765a1b5258779377685c87aaea490c621aed48e508ad19ee8b9b7f" exitCode=0 Oct 02 08:08:56 crc kubenswrapper[4960]: I1002 08:08:56.099542 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" event={"ID":"b169e295-17c4-457d-8832-bb79f85eb5cb","Type":"ContainerDied","Data":"6c6f587982765a1b5258779377685c87aaea490c621aed48e508ad19ee8b9b7f"} Oct 02 08:08:56 crc kubenswrapper[4960]: I1002 08:08:56.110790 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0be16689-f82d-4473-898b-a76cb6885219","Type":"ContainerStarted","Data":"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36"} Oct 02 08:08:56 crc kubenswrapper[4960]: I1002 08:08:56.753699 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.120530 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0be16689-f82d-4473-898b-a76cb6885219","Type":"ContainerStarted","Data":"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b"} Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.120726 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.125414 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"d2695413-bac9-4098-8399-72819bda9615","Type":"ContainerStarted","Data":"30ad267bc1c8360e2537d61a27384fe5b9144dbb9eb8adcef8afd8711a6752de"} Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.128766 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" event={"ID":"b169e295-17c4-457d-8832-bb79f85eb5cb","Type":"ContainerStarted","Data":"5833ebeb1b7c2593cebe27f7e1bd9ab9341def6446f1bd5fd1bec3be1eda28f0"} Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.129146 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.146405 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.146374902 podStartE2EDuration="4.146374902s" podCreationTimestamp="2025-10-02 08:08:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:08:57.13957839 +0000 UTC m=+3158.171524687" watchObservedRunningTime="2025-10-02 08:08:57.146374902 +0000 UTC m=+3158.178321189" Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.175019 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" podStartSLOduration=4.174996528 podStartE2EDuration="4.174996528s" podCreationTimestamp="2025-10-02 08:08:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:08:57.174079713 +0000 UTC m=+3158.206026010" watchObservedRunningTime="2025-10-02 08:08:57.174996528 +0000 UTC m=+3158.206942815" Oct 02 08:08:57 crc kubenswrapper[4960]: I1002 08:08:57.204121 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.495276837 podStartE2EDuration="4.204093196s" podCreationTimestamp="2025-10-02 08:08:53 +0000 UTC" firstStartedPulling="2025-10-02 08:08:54.31726312 +0000 UTC m=+3155.349209407" lastFinishedPulling="2025-10-02 08:08:55.026079479 +0000 UTC m=+3156.058025766" observedRunningTime="2025-10-02 08:08:57.197691635 +0000 UTC m=+3158.229637922" watchObservedRunningTime="2025-10-02 08:08:57.204093196 +0000 UTC m=+3158.236039483" Oct 02 08:08:58 crc kubenswrapper[4960]: I1002 08:08:58.141642 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api-log" containerID="cri-o://9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36" gracePeriod=30 Oct 02 08:08:58 crc kubenswrapper[4960]: I1002 08:08:58.143163 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api" containerID="cri-o://b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b" gracePeriod=30 Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.105288 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.153230 4960 generic.go:334] "Generic (PLEG): container finished" podID="0be16689-f82d-4473-898b-a76cb6885219" containerID="b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b" exitCode=0 Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.153271 4960 generic.go:334] "Generic (PLEG): container finished" podID="0be16689-f82d-4473-898b-a76cb6885219" containerID="9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36" exitCode=143 Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.153300 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0be16689-f82d-4473-898b-a76cb6885219","Type":"ContainerDied","Data":"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b"} Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.153343 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0be16689-f82d-4473-898b-a76cb6885219","Type":"ContainerDied","Data":"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36"} Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.153356 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0be16689-f82d-4473-898b-a76cb6885219","Type":"ContainerDied","Data":"d7d66c51ea1947bdc5d6dd505a450da650d13b923a27b9e97d6f1acfad157a4e"} Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.153357 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.153375 4960 scope.go:117] "RemoveContainer" containerID="b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.204018 4960 scope.go:117] "RemoveContainer" containerID="9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.230486 4960 scope.go:117] "RemoveContainer" containerID="b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b" Oct 02 08:08:59 crc kubenswrapper[4960]: E1002 08:08:59.231068 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b\": container with ID starting with b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b not found: ID does not exist" containerID="b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.231112 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b"} err="failed to get container status \"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b\": rpc error: code = NotFound desc = could not find container \"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b\": container with ID starting with b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b not found: ID does not exist" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.231136 4960 scope.go:117] "RemoveContainer" containerID="9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36" Oct 02 08:08:59 crc kubenswrapper[4960]: E1002 08:08:59.231359 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36\": container with ID starting with 9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36 not found: ID does not exist" containerID="9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.231381 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36"} err="failed to get container status \"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36\": rpc error: code = NotFound desc = could not find container \"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36\": container with ID starting with 9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36 not found: ID does not exist" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.231394 4960 scope.go:117] "RemoveContainer" containerID="b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.231592 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b"} err="failed to get container status \"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b\": rpc error: code = NotFound desc = could not find container \"b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b\": container with ID starting with b99b8b517005edb1f8f8c276aaed7692e36b0744ef12af4aae246e89227b459b not found: ID does not exist" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.231610 4960 scope.go:117] "RemoveContainer" containerID="9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.231782 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36"} err="failed to get container status \"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36\": rpc error: code = NotFound desc = could not find container \"9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36\": container with ID starting with 9c69ac7e942a41e92f3f58b116be7c999db8d8f508de6f513f61957e8ffd3e36 not found: ID does not exist" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.256440 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-scripts\") pod \"0be16689-f82d-4473-898b-a76cb6885219\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.256636 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctb6c\" (UniqueName: \"kubernetes.io/projected/0be16689-f82d-4473-898b-a76cb6885219-kube-api-access-ctb6c\") pod \"0be16689-f82d-4473-898b-a76cb6885219\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.256722 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-combined-ca-bundle\") pod \"0be16689-f82d-4473-898b-a76cb6885219\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.256779 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data-custom\") pod \"0be16689-f82d-4473-898b-a76cb6885219\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.256830 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be16689-f82d-4473-898b-a76cb6885219-logs\") pod \"0be16689-f82d-4473-898b-a76cb6885219\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.256884 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data\") pod \"0be16689-f82d-4473-898b-a76cb6885219\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.257075 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0be16689-f82d-4473-898b-a76cb6885219-etc-machine-id\") pod \"0be16689-f82d-4473-898b-a76cb6885219\" (UID: \"0be16689-f82d-4473-898b-a76cb6885219\") " Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.257554 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0be16689-f82d-4473-898b-a76cb6885219-logs" (OuterVolumeSpecName: "logs") pod "0be16689-f82d-4473-898b-a76cb6885219" (UID: "0be16689-f82d-4473-898b-a76cb6885219"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.257626 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0be16689-f82d-4473-898b-a76cb6885219-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0be16689-f82d-4473-898b-a76cb6885219" (UID: "0be16689-f82d-4473-898b-a76cb6885219"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.264661 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-scripts" (OuterVolumeSpecName: "scripts") pod "0be16689-f82d-4473-898b-a76cb6885219" (UID: "0be16689-f82d-4473-898b-a76cb6885219"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.265103 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0be16689-f82d-4473-898b-a76cb6885219" (UID: "0be16689-f82d-4473-898b-a76cb6885219"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.269415 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0be16689-f82d-4473-898b-a76cb6885219-kube-api-access-ctb6c" (OuterVolumeSpecName: "kube-api-access-ctb6c") pod "0be16689-f82d-4473-898b-a76cb6885219" (UID: "0be16689-f82d-4473-898b-a76cb6885219"). InnerVolumeSpecName "kube-api-access-ctb6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.293843 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0be16689-f82d-4473-898b-a76cb6885219" (UID: "0be16689-f82d-4473-898b-a76cb6885219"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.325853 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data" (OuterVolumeSpecName: "config-data") pod "0be16689-f82d-4473-898b-a76cb6885219" (UID: "0be16689-f82d-4473-898b-a76cb6885219"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.360766 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.360800 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctb6c\" (UniqueName: \"kubernetes.io/projected/0be16689-f82d-4473-898b-a76cb6885219-kube-api-access-ctb6c\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.360814 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.360829 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.360840 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be16689-f82d-4473-898b-a76cb6885219-logs\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.360850 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be16689-f82d-4473-898b-a76cb6885219-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.360861 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0be16689-f82d-4473-898b-a76cb6885219-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.433079 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.433456 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-central-agent" containerID="cri-o://4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1" gracePeriod=30 Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.433507 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="proxy-httpd" containerID="cri-o://9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e" gracePeriod=30 Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.433570 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="sg-core" containerID="cri-o://450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b" gracePeriod=30 Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.433532 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-notification-agent" containerID="cri-o://4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d" gracePeriod=30 Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.496931 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.515106 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.532795 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:59 crc kubenswrapper[4960]: E1002 08:08:59.533598 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.533623 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api" Oct 02 08:08:59 crc kubenswrapper[4960]: E1002 08:08:59.533648 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api-log" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.533656 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api-log" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.533913 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api-log" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.533955 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be16689-f82d-4473-898b-a76cb6885219" containerName="manila-api" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.535214 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.546955 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.547216 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.547377 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.548135 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.679847 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-internal-tls-certs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.679892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-public-tls-certs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.679966 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-scripts\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.680050 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs88z\" (UniqueName: \"kubernetes.io/projected/5773acec-b7dc-4887-b302-832737da62e9-kube-api-access-rs88z\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.680068 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-config-data-custom\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.680084 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5773acec-b7dc-4887-b302-832737da62e9-etc-machine-id\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.680114 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5773acec-b7dc-4887-b302-832737da62e9-logs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.680150 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-config-data\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.680164 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.782892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-scripts\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783049 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs88z\" (UniqueName: \"kubernetes.io/projected/5773acec-b7dc-4887-b302-832737da62e9-kube-api-access-rs88z\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783079 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5773acec-b7dc-4887-b302-832737da62e9-etc-machine-id\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783106 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-config-data-custom\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783145 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5773acec-b7dc-4887-b302-832737da62e9-logs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783198 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-config-data\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783224 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783216 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5773acec-b7dc-4887-b302-832737da62e9-etc-machine-id\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.783648 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5773acec-b7dc-4887-b302-832737da62e9-logs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.784730 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-internal-tls-certs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.784811 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-public-tls-certs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.790433 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-scripts\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.794616 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-internal-tls-certs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.794634 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-public-tls-certs\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.794787 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.794986 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-config-data-custom\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.795316 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5773acec-b7dc-4887-b302-832737da62e9-config-data\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.806222 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs88z\" (UniqueName: \"kubernetes.io/projected/5773acec-b7dc-4887-b302-832737da62e9-kube-api-access-rs88z\") pod \"manila-api-0\" (UID: \"5773acec-b7dc-4887-b302-832737da62e9\") " pod="openstack/manila-api-0" Oct 02 08:08:59 crc kubenswrapper[4960]: I1002 08:08:59.902843 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 08:09:00 crc kubenswrapper[4960]: I1002 08:09:00.166708 4960 generic.go:334] "Generic (PLEG): container finished" podID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerID="9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e" exitCode=0 Oct 02 08:09:00 crc kubenswrapper[4960]: I1002 08:09:00.166749 4960 generic.go:334] "Generic (PLEG): container finished" podID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerID="450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b" exitCode=2 Oct 02 08:09:00 crc kubenswrapper[4960]: I1002 08:09:00.166758 4960 generic.go:334] "Generic (PLEG): container finished" podID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerID="4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1" exitCode=0 Oct 02 08:09:00 crc kubenswrapper[4960]: I1002 08:09:00.166787 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerDied","Data":"9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e"} Oct 02 08:09:00 crc kubenswrapper[4960]: I1002 08:09:00.166875 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerDied","Data":"450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b"} Oct 02 08:09:00 crc kubenswrapper[4960]: I1002 08:09:00.166889 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerDied","Data":"4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1"} Oct 02 08:09:00 crc kubenswrapper[4960]: I1002 08:09:00.344778 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0be16689-f82d-4473-898b-a76cb6885219" path="/var/lib/kubelet/pods/0be16689-f82d-4473-898b-a76cb6885219/volumes" Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.330215 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:09:02 crc kubenswrapper[4960]: E1002 08:09:02.330879 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.873685 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.973769 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-ceilometer-tls-certs\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.973822 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-log-httpd\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.973861 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.974605 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.974712 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-scripts\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.975178 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-run-httpd\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.975259 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-config-data\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.975292 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-sg-core-conf-yaml\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.975309 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5ndw\" (UniqueName: \"kubernetes.io/projected/967ca6d6-7b63-4683-b829-fa7e36d23cc6-kube-api-access-q5ndw\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.975652 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.976170 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.982099 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/967ca6d6-7b63-4683-b829-fa7e36d23cc6-kube-api-access-q5ndw" (OuterVolumeSpecName: "kube-api-access-q5ndw") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "kube-api-access-q5ndw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:09:02 crc kubenswrapper[4960]: I1002 08:09:02.984464 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-scripts" (OuterVolumeSpecName: "scripts") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.019306 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.037079 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.076365 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.077407 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle\") pod \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\" (UID: \"967ca6d6-7b63-4683-b829-fa7e36d23cc6\") " Oct 02 08:09:03 crc kubenswrapper[4960]: W1002 08:09:03.077817 4960 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/967ca6d6-7b63-4683-b829-fa7e36d23cc6/volumes/kubernetes.io~secret/combined-ca-bundle Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.077952 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.078374 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/967ca6d6-7b63-4683-b829-fa7e36d23cc6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.078400 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.078415 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5ndw\" (UniqueName: \"kubernetes.io/projected/967ca6d6-7b63-4683-b829-fa7e36d23cc6-kube-api-access-q5ndw\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.078424 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.078432 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.078446 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.104831 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-config-data" (OuterVolumeSpecName: "config-data") pod "967ca6d6-7b63-4683-b829-fa7e36d23cc6" (UID: "967ca6d6-7b63-4683-b829-fa7e36d23cc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.154950 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.180597 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/967ca6d6-7b63-4683-b829-fa7e36d23cc6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.210267 4960 generic.go:334] "Generic (PLEG): container finished" podID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerID="4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d" exitCode=0 Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.210345 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerDied","Data":"4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d"} Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.210386 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"967ca6d6-7b63-4683-b829-fa7e36d23cc6","Type":"ContainerDied","Data":"b8428fd482a7b0fc9dad3a24bc444ed126410c796a87ad37c37e11b739c6e837"} Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.210407 4960 scope.go:117] "RemoveContainer" containerID="9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.210615 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.218435 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5773acec-b7dc-4887-b302-832737da62e9","Type":"ContainerStarted","Data":"bc13227c3abc12520eb9c0c44e132c1ebece71739fb507b88cd369ff70586a9a"} Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.253908 4960 scope.go:117] "RemoveContainer" containerID="450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.264664 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.278417 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.297644 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.298530 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="proxy-httpd" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298552 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="proxy-httpd" Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.298582 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-central-agent" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298589 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-central-agent" Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.298607 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="sg-core" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298613 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="sg-core" Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.298635 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-notification-agent" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298641 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-notification-agent" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298825 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="proxy-httpd" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298843 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="sg-core" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298853 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-notification-agent" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.298863 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" containerName="ceilometer-central-agent" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.302390 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.309460 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.318000 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.318046 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.318289 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.326796 4960 scope.go:117] "RemoveContainer" containerID="4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.384959 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-run-httpd\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385172 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385208 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-config-data\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385229 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-log-httpd\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385274 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385306 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-scripts\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385391 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lflb\" (UniqueName: \"kubernetes.io/projected/e78f966b-6879-4fae-ae4e-660aa67fb119-kube-api-access-7lflb\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.385615 4960 scope.go:117] "RemoveContainer" containerID="4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.419727 4960 scope.go:117] "RemoveContainer" containerID="9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e" Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.420591 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e\": container with ID starting with 9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e not found: ID does not exist" containerID="9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.420633 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e"} err="failed to get container status \"9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e\": rpc error: code = NotFound desc = could not find container \"9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e\": container with ID starting with 9cbd2abf40b245d565c4c02bc38e4cc6f6df878ffe80cb1c2dee39a45a64ef9e not found: ID does not exist" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.420664 4960 scope.go:117] "RemoveContainer" containerID="450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b" Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.421193 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b\": container with ID starting with 450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b not found: ID does not exist" containerID="450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.421250 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b"} err="failed to get container status \"450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b\": rpc error: code = NotFound desc = could not find container \"450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b\": container with ID starting with 450555d5e865ff7bf2f5feb0aecb49312c7f03c4dde306ebce72f29e435cdf1b not found: ID does not exist" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.421289 4960 scope.go:117] "RemoveContainer" containerID="4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d" Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.421843 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d\": container with ID starting with 4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d not found: ID does not exist" containerID="4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.421907 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d"} err="failed to get container status \"4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d\": rpc error: code = NotFound desc = could not find container \"4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d\": container with ID starting with 4e56489e52766782083e05d1a0d6829f200f77739c80cbd0cf6791166caad65d not found: ID does not exist" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.421956 4960 scope.go:117] "RemoveContainer" containerID="4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1" Oct 02 08:09:03 crc kubenswrapper[4960]: E1002 08:09:03.422495 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1\": container with ID starting with 4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1 not found: ID does not exist" containerID="4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.422527 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1"} err="failed to get container status \"4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1\": rpc error: code = NotFound desc = could not find container \"4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1\": container with ID starting with 4d51d49d43e58708de216ceca5de09b5573e1a67722b8e2a4292c58ce3187db1 not found: ID does not exist" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.489999 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-run-httpd\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.490127 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.490150 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-config-data\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.490167 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-log-httpd\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.490216 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.490247 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-scripts\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.490293 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.490325 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lflb\" (UniqueName: \"kubernetes.io/projected/e78f966b-6879-4fae-ae4e-660aa67fb119-kube-api-access-7lflb\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.491340 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-run-httpd\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.492528 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-log-httpd\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.497884 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.498169 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-config-data\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.498967 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-scripts\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.500123 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.501152 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.512772 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lflb\" (UniqueName: \"kubernetes.io/projected/e78f966b-6879-4fae-ae4e-660aa67fb119-kube-api-access-7lflb\") pod \"ceilometer-0\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.639450 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.704237 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.888288 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6785d65565-vfxf8" Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.970653 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b897d8569-xzwfc"] Oct 02 08:09:03 crc kubenswrapper[4960]: I1002 08:09:03.971318 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" podUID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerName="dnsmasq-dns" containerID="cri-o://94627eaaf9bb69e5565cec1e0d2fdf8a94a0592615e83f84adb2c84da8f1a2cf" gracePeriod=10 Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.250610 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.280732 4960 generic.go:334] "Generic (PLEG): container finished" podID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerID="94627eaaf9bb69e5565cec1e0d2fdf8a94a0592615e83f84adb2c84da8f1a2cf" exitCode=0 Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.281220 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" event={"ID":"17b9b7cb-39cd-45fd-9653-fe88daede51a","Type":"ContainerDied","Data":"94627eaaf9bb69e5565cec1e0d2fdf8a94a0592615e83f84adb2c84da8f1a2cf"} Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.289823 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5773acec-b7dc-4887-b302-832737da62e9","Type":"ContainerStarted","Data":"d6ece137cdf6e422c2bcc21d5e13ef3aa42581dce4035bde72eed1f820eaa3cc"} Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.292964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ad1f362d-1d99-4948-a74e-cb656fbc3825","Type":"ContainerStarted","Data":"a9ed8b134a23a0ccef28e6585652bec0710917d13c7bce87b2409049c33594db"} Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.293014 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ad1f362d-1d99-4948-a74e-cb656fbc3825","Type":"ContainerStarted","Data":"ba2cfc4f3f33a09417513f041a1d9069be25c54b809775aa0436ebc4a1d2f566"} Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.330023 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.539550321 podStartE2EDuration="11.329965959s" podCreationTimestamp="2025-10-02 08:08:53 +0000 UTC" firstStartedPulling="2025-10-02 08:08:54.813746369 +0000 UTC m=+3155.845692656" lastFinishedPulling="2025-10-02 08:09:02.604162017 +0000 UTC m=+3163.636108294" observedRunningTime="2025-10-02 08:09:04.325352276 +0000 UTC m=+3165.357298573" watchObservedRunningTime="2025-10-02 08:09:04.329965959 +0000 UTC m=+3165.361912246" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.367519 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="967ca6d6-7b63-4683-b829-fa7e36d23cc6" path="/var/lib/kubelet/pods/967ca6d6-7b63-4683-b829-fa7e36d23cc6/volumes" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.623142 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.630276 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-dns-svc\") pod \"17b9b7cb-39cd-45fd-9653-fe88daede51a\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.630364 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdpfs\" (UniqueName: \"kubernetes.io/projected/17b9b7cb-39cd-45fd-9653-fe88daede51a-kube-api-access-pdpfs\") pod \"17b9b7cb-39cd-45fd-9653-fe88daede51a\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.630574 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-sb\") pod \"17b9b7cb-39cd-45fd-9653-fe88daede51a\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.630810 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-config\") pod \"17b9b7cb-39cd-45fd-9653-fe88daede51a\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.630841 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-nb\") pod \"17b9b7cb-39cd-45fd-9653-fe88daede51a\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.630868 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-openstack-edpm-ipam\") pod \"17b9b7cb-39cd-45fd-9653-fe88daede51a\" (UID: \"17b9b7cb-39cd-45fd-9653-fe88daede51a\") " Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.638663 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17b9b7cb-39cd-45fd-9653-fe88daede51a-kube-api-access-pdpfs" (OuterVolumeSpecName: "kube-api-access-pdpfs") pod "17b9b7cb-39cd-45fd-9653-fe88daede51a" (UID: "17b9b7cb-39cd-45fd-9653-fe88daede51a"). InnerVolumeSpecName "kube-api-access-pdpfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.725862 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17b9b7cb-39cd-45fd-9653-fe88daede51a" (UID: "17b9b7cb-39cd-45fd-9653-fe88daede51a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.734055 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdpfs\" (UniqueName: \"kubernetes.io/projected/17b9b7cb-39cd-45fd-9653-fe88daede51a-kube-api-access-pdpfs\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.734081 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.741633 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-config" (OuterVolumeSpecName: "config") pod "17b9b7cb-39cd-45fd-9653-fe88daede51a" (UID: "17b9b7cb-39cd-45fd-9653-fe88daede51a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.769654 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "17b9b7cb-39cd-45fd-9653-fe88daede51a" (UID: "17b9b7cb-39cd-45fd-9653-fe88daede51a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.770626 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17b9b7cb-39cd-45fd-9653-fe88daede51a" (UID: "17b9b7cb-39cd-45fd-9653-fe88daede51a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.780465 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17b9b7cb-39cd-45fd-9653-fe88daede51a" (UID: "17b9b7cb-39cd-45fd-9653-fe88daede51a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.836954 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-config\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.837318 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.837331 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:04 crc kubenswrapper[4960]: I1002 08:09:04.837344 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17b9b7cb-39cd-45fd-9653-fe88daede51a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.310178 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5773acec-b7dc-4887-b302-832737da62e9","Type":"ContainerStarted","Data":"193f683a82986c9ef566f109d4962d68dc07449500bb92b14403c174f9b31a22"} Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.310468 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.312755 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerStarted","Data":"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f"} Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.312804 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerStarted","Data":"42a2291a86f5f009952cf4cb3a75be8490f8b2f9d9ba6840cd6585d4634d77f8"} Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.317352 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" event={"ID":"17b9b7cb-39cd-45fd-9653-fe88daede51a","Type":"ContainerDied","Data":"2bb83c30931c6c2260b1b99b1d53b6f5158b1b65725b1c94dbc1aab0fe8a11ac"} Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.317416 4960 scope.go:117] "RemoveContainer" containerID="94627eaaf9bb69e5565cec1e0d2fdf8a94a0592615e83f84adb2c84da8f1a2cf" Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.317922 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b897d8569-xzwfc" Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.360735 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=6.360706859 podStartE2EDuration="6.360706859s" podCreationTimestamp="2025-10-02 08:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:09:05.352476198 +0000 UTC m=+3166.384422485" watchObservedRunningTime="2025-10-02 08:09:05.360706859 +0000 UTC m=+3166.392653146" Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.372599 4960 scope.go:117] "RemoveContainer" containerID="2606b40635861e61bd9ea79b4756136ec706e29498af737a07d7d3d090bfc0ed" Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.412102 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b897d8569-xzwfc"] Oct 02 08:09:05 crc kubenswrapper[4960]: I1002 08:09:05.427099 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b897d8569-xzwfc"] Oct 02 08:09:06 crc kubenswrapper[4960]: I1002 08:09:06.344588 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17b9b7cb-39cd-45fd-9653-fe88daede51a" path="/var/lib/kubelet/pods/17b9b7cb-39cd-45fd-9653-fe88daede51a/volumes" Oct 02 08:09:06 crc kubenswrapper[4960]: I1002 08:09:06.346048 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerStarted","Data":"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6"} Oct 02 08:09:07 crc kubenswrapper[4960]: I1002 08:09:07.351765 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerStarted","Data":"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3"} Oct 02 08:09:07 crc kubenswrapper[4960]: I1002 08:09:07.389947 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:09 crc kubenswrapper[4960]: I1002 08:09:09.401625 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerStarted","Data":"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43"} Oct 02 08:09:09 crc kubenswrapper[4960]: I1002 08:09:09.402585 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-central-agent" containerID="cri-o://6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" gracePeriod=30 Oct 02 08:09:09 crc kubenswrapper[4960]: I1002 08:09:09.403026 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 08:09:09 crc kubenswrapper[4960]: I1002 08:09:09.403085 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="proxy-httpd" containerID="cri-o://96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" gracePeriod=30 Oct 02 08:09:09 crc kubenswrapper[4960]: I1002 08:09:09.403168 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="sg-core" containerID="cri-o://3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" gracePeriod=30 Oct 02 08:09:09 crc kubenswrapper[4960]: I1002 08:09:09.403244 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-notification-agent" containerID="cri-o://fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" gracePeriod=30 Oct 02 08:09:09 crc kubenswrapper[4960]: I1002 08:09:09.443738 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.519924787 podStartE2EDuration="6.44371339s" podCreationTimestamp="2025-10-02 08:09:03 +0000 UTC" firstStartedPulling="2025-10-02 08:09:04.257295805 +0000 UTC m=+3165.289242092" lastFinishedPulling="2025-10-02 08:09:08.181084398 +0000 UTC m=+3169.213030695" observedRunningTime="2025-10-02 08:09:09.429315976 +0000 UTC m=+3170.461262303" watchObservedRunningTime="2025-10-02 08:09:09.44371339 +0000 UTC m=+3170.475659687" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.248056 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412367 4960 generic.go:334] "Generic (PLEG): container finished" podID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerID="96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" exitCode=0 Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412403 4960 generic.go:334] "Generic (PLEG): container finished" podID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerID="3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" exitCode=2 Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412407 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412443 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerDied","Data":"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43"} Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412484 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerDied","Data":"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3"} Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412515 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerDied","Data":"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6"} Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412534 4960 generic.go:334] "Generic (PLEG): container finished" podID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerID="fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" exitCode=0 Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412550 4960 generic.go:334] "Generic (PLEG): container finished" podID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerID="6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" exitCode=0 Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412529 4960 scope.go:117] "RemoveContainer" containerID="96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerDied","Data":"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f"} Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.412607 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e78f966b-6879-4fae-ae4e-660aa67fb119","Type":"ContainerDied","Data":"42a2291a86f5f009952cf4cb3a75be8490f8b2f9d9ba6840cd6585d4634d77f8"} Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.419457 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-ceilometer-tls-certs\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.419554 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-scripts\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.419617 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-log-httpd\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.419809 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-sg-core-conf-yaml\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.419860 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-run-httpd\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.419888 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-config-data\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.419953 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lflb\" (UniqueName: \"kubernetes.io/projected/e78f966b-6879-4fae-ae4e-660aa67fb119-kube-api-access-7lflb\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.420246 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.420267 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.420289 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-combined-ca-bundle\") pod \"e78f966b-6879-4fae-ae4e-660aa67fb119\" (UID: \"e78f966b-6879-4fae-ae4e-660aa67fb119\") " Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.421884 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.421916 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e78f966b-6879-4fae-ae4e-660aa67fb119-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.426687 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-scripts" (OuterVolumeSpecName: "scripts") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.426826 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78f966b-6879-4fae-ae4e-660aa67fb119-kube-api-access-7lflb" (OuterVolumeSpecName: "kube-api-access-7lflb") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "kube-api-access-7lflb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.447284 4960 scope.go:117] "RemoveContainer" containerID="3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.448677 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.468520 4960 scope.go:117] "RemoveContainer" containerID="fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.476840 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.491229 4960 scope.go:117] "RemoveContainer" containerID="6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.497598 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.512833 4960 scope.go:117] "RemoveContainer" containerID="96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.513306 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": container with ID starting with 96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43 not found: ID does not exist" containerID="96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.513348 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43"} err="failed to get container status \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": rpc error: code = NotFound desc = could not find container \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": container with ID starting with 96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.513376 4960 scope.go:117] "RemoveContainer" containerID="3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.513745 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": container with ID starting with 3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3 not found: ID does not exist" containerID="3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.513773 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3"} err="failed to get container status \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": rpc error: code = NotFound desc = could not find container \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": container with ID starting with 3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.513790 4960 scope.go:117] "RemoveContainer" containerID="fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.514106 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": container with ID starting with fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6 not found: ID does not exist" containerID="fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.514162 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6"} err="failed to get container status \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": rpc error: code = NotFound desc = could not find container \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": container with ID starting with fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.514202 4960 scope.go:117] "RemoveContainer" containerID="6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.514589 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": container with ID starting with 6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f not found: ID does not exist" containerID="6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.514632 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f"} err="failed to get container status \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": rpc error: code = NotFound desc = could not find container \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": container with ID starting with 6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.514647 4960 scope.go:117] "RemoveContainer" containerID="96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.514903 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43"} err="failed to get container status \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": rpc error: code = NotFound desc = could not find container \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": container with ID starting with 96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.514932 4960 scope.go:117] "RemoveContainer" containerID="3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.515302 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3"} err="failed to get container status \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": rpc error: code = NotFound desc = could not find container \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": container with ID starting with 3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.515337 4960 scope.go:117] "RemoveContainer" containerID="fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.515608 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6"} err="failed to get container status \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": rpc error: code = NotFound desc = could not find container \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": container with ID starting with fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.515644 4960 scope.go:117] "RemoveContainer" containerID="6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.516139 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f"} err="failed to get container status \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": rpc error: code = NotFound desc = could not find container \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": container with ID starting with 6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.516162 4960 scope.go:117] "RemoveContainer" containerID="96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.516496 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43"} err="failed to get container status \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": rpc error: code = NotFound desc = could not find container \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": container with ID starting with 96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.516523 4960 scope.go:117] "RemoveContainer" containerID="3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.516777 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3"} err="failed to get container status \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": rpc error: code = NotFound desc = could not find container \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": container with ID starting with 3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.516799 4960 scope.go:117] "RemoveContainer" containerID="fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.517107 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6"} err="failed to get container status \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": rpc error: code = NotFound desc = could not find container \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": container with ID starting with fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.517131 4960 scope.go:117] "RemoveContainer" containerID="6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.517396 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f"} err="failed to get container status \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": rpc error: code = NotFound desc = could not find container \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": container with ID starting with 6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.517420 4960 scope.go:117] "RemoveContainer" containerID="96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.517659 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43"} err="failed to get container status \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": rpc error: code = NotFound desc = could not find container \"96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43\": container with ID starting with 96e1efe918dcf0b7e026ec32a06c1cb91ec7f7f6458731a52086ab4a861b5a43 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.517682 4960 scope.go:117] "RemoveContainer" containerID="3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.518081 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3"} err="failed to get container status \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": rpc error: code = NotFound desc = could not find container \"3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3\": container with ID starting with 3ba6961bb3dda6fd30e3bf41e64d5cb8d49ab59dfa59fd658419d95fbec28ae3 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.518100 4960 scope.go:117] "RemoveContainer" containerID="fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.518286 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6"} err="failed to get container status \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": rpc error: code = NotFound desc = could not find container \"fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6\": container with ID starting with fce22fac8a43dadccb0f4e9bf6488ad0ba17b66e8ca4e03f991339447f55d7e6 not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.518303 4960 scope.go:117] "RemoveContainer" containerID="6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.518573 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f"} err="failed to get container status \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": rpc error: code = NotFound desc = could not find container \"6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f\": container with ID starting with 6b5fb9ef890837ed584dcf56d7b6eb06053b4f2e9cec29d6ab0e96b857f1b43f not found: ID does not exist" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.524612 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.524635 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.524647 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.524656 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.524668 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lflb\" (UniqueName: \"kubernetes.io/projected/e78f966b-6879-4fae-ae4e-660aa67fb119-kube-api-access-7lflb\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.531006 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-config-data" (OuterVolumeSpecName: "config-data") pod "e78f966b-6879-4fae-ae4e-660aa67fb119" (UID: "e78f966b-6879-4fae-ae4e-660aa67fb119"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.626797 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e78f966b-6879-4fae-ae4e-660aa67fb119-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.748770 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.759399 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.809612 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.810674 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerName="dnsmasq-dns" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.810701 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerName="dnsmasq-dns" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.810764 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="proxy-httpd" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.810776 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="proxy-httpd" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.810835 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="sg-core" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.810844 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="sg-core" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.810868 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerName="init" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.810877 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerName="init" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.810908 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-notification-agent" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.810918 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-notification-agent" Oct 02 08:09:10 crc kubenswrapper[4960]: E1002 08:09:10.810945 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-central-agent" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.810957 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-central-agent" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.811487 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-central-agent" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.811527 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="proxy-httpd" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.811563 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="ceilometer-notification-agent" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.811586 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="17b9b7cb-39cd-45fd-9653-fe88daede51a" containerName="dnsmasq-dns" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.811603 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" containerName="sg-core" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.820099 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.820259 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.823139 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.823499 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.824787 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932384 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932457 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-scripts\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932486 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932620 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-log-httpd\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932733 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932757 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf2xv\" (UniqueName: \"kubernetes.io/projected/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-kube-api-access-jf2xv\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932781 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-config-data\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:10 crc kubenswrapper[4960]: I1002 08:09:10.932808 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-run-httpd\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035044 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-log-httpd\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035466 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035486 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf2xv\" (UniqueName: \"kubernetes.io/projected/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-kube-api-access-jf2xv\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035504 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-config-data\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035525 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-run-httpd\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035556 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035580 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-scripts\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035595 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.035591 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-log-httpd\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.036167 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-run-httpd\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.039548 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-scripts\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.040518 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.042408 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.047552 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.054277 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-config-data\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.063489 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf2xv\" (UniqueName: \"kubernetes.io/projected/d1225542-cda6-49d9-bdf3-6bbb7d17e8ac-kube-api-access-jf2xv\") pod \"ceilometer-0\" (UID: \"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac\") " pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.144492 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 08:09:11 crc kubenswrapper[4960]: I1002 08:09:11.667632 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 08:09:11 crc kubenswrapper[4960]: W1002 08:09:11.673522 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1225542_cda6_49d9_bdf3_6bbb7d17e8ac.slice/crio-c62f5c73a8b876c575b0b8a0506cef4fa61a4ff40229738638ab272affe6603d WatchSource:0}: Error finding container c62f5c73a8b876c575b0b8a0506cef4fa61a4ff40229738638ab272affe6603d: Status 404 returned error can't find the container with id c62f5c73a8b876c575b0b8a0506cef4fa61a4ff40229738638ab272affe6603d Oct 02 08:09:12 crc kubenswrapper[4960]: I1002 08:09:12.351780 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78f966b-6879-4fae-ae4e-660aa67fb119" path="/var/lib/kubelet/pods/e78f966b-6879-4fae-ae4e-660aa67fb119/volumes" Oct 02 08:09:12 crc kubenswrapper[4960]: I1002 08:09:12.443326 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac","Type":"ContainerStarted","Data":"29c29dc883bbb16920ba45183a98517483ebf252a2a0620a1b3c0831b8431f8c"} Oct 02 08:09:12 crc kubenswrapper[4960]: I1002 08:09:12.443378 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac","Type":"ContainerStarted","Data":"c62f5c73a8b876c575b0b8a0506cef4fa61a4ff40229738638ab272affe6603d"} Oct 02 08:09:13 crc kubenswrapper[4960]: I1002 08:09:13.466595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac","Type":"ContainerStarted","Data":"5e03c831cf8c961399d6e31ce6e84e0d411576c91b268e2990c25a1749d99c7d"} Oct 02 08:09:13 crc kubenswrapper[4960]: I1002 08:09:13.797589 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 02 08:09:14 crc kubenswrapper[4960]: I1002 08:09:14.483911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac","Type":"ContainerStarted","Data":"b032cf13ee474a6197594d2715540bed3cfd62d9004a5f0201a3462fe8c43cbd"} Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.330800 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:09:15 crc kubenswrapper[4960]: E1002 08:09:15.331787 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.339588 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.387038 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.492089 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.531258 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="manila-scheduler" containerID="cri-o://083cc1cd63078041bcb54566f3f1f0878218809774cf45d62c9d1792520b4aa3" gracePeriod=30 Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.532108 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1225542-cda6-49d9-bdf3-6bbb7d17e8ac","Type":"ContainerStarted","Data":"0b0d039ca48b6023f85c65a7e1a4c5c3769bcfcda8ff5335d1f7c4683496b07d"} Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.532236 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="probe" containerID="cri-o://30ad267bc1c8360e2537d61a27384fe5b9144dbb9eb8adcef8afd8711a6752de" gracePeriod=30 Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.532332 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.615042 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:09:15 crc kubenswrapper[4960]: I1002 08:09:15.617938 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.475519066 podStartE2EDuration="5.617913229s" podCreationTimestamp="2025-10-02 08:09:10 +0000 UTC" firstStartedPulling="2025-10-02 08:09:11.676466802 +0000 UTC m=+3172.708413089" lastFinishedPulling="2025-10-02 08:09:14.818860915 +0000 UTC m=+3175.850807252" observedRunningTime="2025-10-02 08:09:15.608712002 +0000 UTC m=+3176.640658289" watchObservedRunningTime="2025-10-02 08:09:15.617913229 +0000 UTC m=+3176.649859516" Oct 02 08:09:16 crc kubenswrapper[4960]: I1002 08:09:16.542860 4960 generic.go:334] "Generic (PLEG): container finished" podID="d2695413-bac9-4098-8399-72819bda9615" containerID="30ad267bc1c8360e2537d61a27384fe5b9144dbb9eb8adcef8afd8711a6752de" exitCode=0 Oct 02 08:09:16 crc kubenswrapper[4960]: I1002 08:09:16.542995 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"d2695413-bac9-4098-8399-72819bda9615","Type":"ContainerDied","Data":"30ad267bc1c8360e2537d61a27384fe5b9144dbb9eb8adcef8afd8711a6752de"} Oct 02 08:09:16 crc kubenswrapper[4960]: I1002 08:09:16.543726 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="manila-share" containerID="cri-o://ba2cfc4f3f33a09417513f041a1d9069be25c54b809775aa0436ebc4a1d2f566" gracePeriod=30 Oct 02 08:09:16 crc kubenswrapper[4960]: I1002 08:09:16.543817 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="probe" containerID="cri-o://a9ed8b134a23a0ccef28e6585652bec0710917d13c7bce87b2409049c33594db" gracePeriod=30 Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.556441 4960 generic.go:334] "Generic (PLEG): container finished" podID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerID="a9ed8b134a23a0ccef28e6585652bec0710917d13c7bce87b2409049c33594db" exitCode=0 Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.557278 4960 generic.go:334] "Generic (PLEG): container finished" podID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerID="ba2cfc4f3f33a09417513f041a1d9069be25c54b809775aa0436ebc4a1d2f566" exitCode=1 Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.556524 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ad1f362d-1d99-4948-a74e-cb656fbc3825","Type":"ContainerDied","Data":"a9ed8b134a23a0ccef28e6585652bec0710917d13c7bce87b2409049c33594db"} Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.557332 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ad1f362d-1d99-4948-a74e-cb656fbc3825","Type":"ContainerDied","Data":"ba2cfc4f3f33a09417513f041a1d9069be25c54b809775aa0436ebc4a1d2f566"} Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.557349 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ad1f362d-1d99-4948-a74e-cb656fbc3825","Type":"ContainerDied","Data":"0fa97d936f32caf1b20877374200e634589d2e3e797867125a24c7ec8c502fd2"} Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.557362 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fa97d936f32caf1b20877374200e634589d2e3e797867125a24c7ec8c502fd2" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.612583 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814134 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-scripts\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814272 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-ceph\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814311 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-combined-ca-bundle\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814416 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dsql\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-kube-api-access-6dsql\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814511 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814545 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-var-lib-manila\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814699 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data-custom\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.814729 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-etc-machine-id\") pod \"ad1f362d-1d99-4948-a74e-cb656fbc3825\" (UID: \"ad1f362d-1d99-4948-a74e-cb656fbc3825\") " Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.816494 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.820078 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.827528 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-scripts" (OuterVolumeSpecName: "scripts") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.827803 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-ceph" (OuterVolumeSpecName: "ceph") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.827949 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-kube-api-access-6dsql" (OuterVolumeSpecName: "kube-api-access-6dsql") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "kube-api-access-6dsql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.831271 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.880028 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.919251 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dsql\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-kube-api-access-6dsql\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.919294 4960 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.919308 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.919322 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad1f362d-1d99-4948-a74e-cb656fbc3825-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.919335 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.919350 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ad1f362d-1d99-4948-a74e-cb656fbc3825-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.919364 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:17 crc kubenswrapper[4960]: I1002 08:09:17.967913 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data" (OuterVolumeSpecName: "config-data") pod "ad1f362d-1d99-4948-a74e-cb656fbc3825" (UID: "ad1f362d-1d99-4948-a74e-cb656fbc3825"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.022901 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad1f362d-1d99-4948-a74e-cb656fbc3825-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.572060 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.609880 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.630047 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.639793 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:09:18 crc kubenswrapper[4960]: E1002 08:09:18.640379 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="probe" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.640405 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="probe" Oct 02 08:09:18 crc kubenswrapper[4960]: E1002 08:09:18.640419 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="manila-share" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.640426 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="manila-share" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.640677 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="probe" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.640696 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" containerName="manila-share" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.641836 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.644209 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.653763 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.743835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-config-data\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.743912 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/29e1c556-f157-4c20-93a7-d0ceecc98439-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.744018 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-scripts\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.744058 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29e1c556-f157-4c20-93a7-d0ceecc98439-ceph\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.744120 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl9cr\" (UniqueName: \"kubernetes.io/projected/29e1c556-f157-4c20-93a7-d0ceecc98439-kube-api-access-gl9cr\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.744489 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29e1c556-f157-4c20-93a7-d0ceecc98439-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.744539 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.744570 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846329 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl9cr\" (UniqueName: \"kubernetes.io/projected/29e1c556-f157-4c20-93a7-d0ceecc98439-kube-api-access-gl9cr\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846461 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29e1c556-f157-4c20-93a7-d0ceecc98439-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846491 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846513 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846544 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-config-data\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846585 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/29e1c556-f157-4c20-93a7-d0ceecc98439-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846630 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29e1c556-f157-4c20-93a7-d0ceecc98439-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846638 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-scripts\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846815 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29e1c556-f157-4c20-93a7-d0ceecc98439-ceph\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.846924 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/29e1c556-f157-4c20-93a7-d0ceecc98439-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.851204 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.851481 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-config-data\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.851660 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-scripts\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.851804 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e1c556-f157-4c20-93a7-d0ceecc98439-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.855710 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29e1c556-f157-4c20-93a7-d0ceecc98439-ceph\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.867184 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl9cr\" (UniqueName: \"kubernetes.io/projected/29e1c556-f157-4c20-93a7-d0ceecc98439-kube-api-access-gl9cr\") pod \"manila-share-share1-0\" (UID: \"29e1c556-f157-4c20-93a7-d0ceecc98439\") " pod="openstack/manila-share-share1-0" Oct 02 08:09:18 crc kubenswrapper[4960]: I1002 08:09:18.993873 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.548742 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 08:09:19 crc kubenswrapper[4960]: W1002 08:09:19.562582 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29e1c556_f157_4c20_93a7_d0ceecc98439.slice/crio-0d3cc006af54f98f2b5d2b101ea030c3f4d74753bc84fd5461942afade790536 WatchSource:0}: Error finding container 0d3cc006af54f98f2b5d2b101ea030c3f4d74753bc84fd5461942afade790536: Status 404 returned error can't find the container with id 0d3cc006af54f98f2b5d2b101ea030c3f4d74753bc84fd5461942afade790536 Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.586730 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"29e1c556-f157-4c20-93a7-d0ceecc98439","Type":"ContainerStarted","Data":"0d3cc006af54f98f2b5d2b101ea030c3f4d74753bc84fd5461942afade790536"} Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.590291 4960 generic.go:334] "Generic (PLEG): container finished" podID="d2695413-bac9-4098-8399-72819bda9615" containerID="083cc1cd63078041bcb54566f3f1f0878218809774cf45d62c9d1792520b4aa3" exitCode=0 Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.590332 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"d2695413-bac9-4098-8399-72819bda9615","Type":"ContainerDied","Data":"083cc1cd63078041bcb54566f3f1f0878218809774cf45d62c9d1792520b4aa3"} Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.590353 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"d2695413-bac9-4098-8399-72819bda9615","Type":"ContainerDied","Data":"869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8"} Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.590371 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.671541 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.764785 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-scripts\") pod \"d2695413-bac9-4098-8399-72819bda9615\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.765092 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r57wf\" (UniqueName: \"kubernetes.io/projected/d2695413-bac9-4098-8399-72819bda9615-kube-api-access-r57wf\") pod \"d2695413-bac9-4098-8399-72819bda9615\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.765158 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data-custom\") pod \"d2695413-bac9-4098-8399-72819bda9615\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.765302 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-combined-ca-bundle\") pod \"d2695413-bac9-4098-8399-72819bda9615\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.765434 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data\") pod \"d2695413-bac9-4098-8399-72819bda9615\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.765463 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2695413-bac9-4098-8399-72819bda9615-etc-machine-id\") pod \"d2695413-bac9-4098-8399-72819bda9615\" (UID: \"d2695413-bac9-4098-8399-72819bda9615\") " Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.766139 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2695413-bac9-4098-8399-72819bda9615-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d2695413-bac9-4098-8399-72819bda9615" (UID: "d2695413-bac9-4098-8399-72819bda9615"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.766771 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2695413-bac9-4098-8399-72819bda9615-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.769667 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-scripts" (OuterVolumeSpecName: "scripts") pod "d2695413-bac9-4098-8399-72819bda9615" (UID: "d2695413-bac9-4098-8399-72819bda9615"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.770088 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2695413-bac9-4098-8399-72819bda9615-kube-api-access-r57wf" (OuterVolumeSpecName: "kube-api-access-r57wf") pod "d2695413-bac9-4098-8399-72819bda9615" (UID: "d2695413-bac9-4098-8399-72819bda9615"). InnerVolumeSpecName "kube-api-access-r57wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.771438 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d2695413-bac9-4098-8399-72819bda9615" (UID: "d2695413-bac9-4098-8399-72819bda9615"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.834247 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2695413-bac9-4098-8399-72819bda9615" (UID: "d2695413-bac9-4098-8399-72819bda9615"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.869239 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r57wf\" (UniqueName: \"kubernetes.io/projected/d2695413-bac9-4098-8399-72819bda9615-kube-api-access-r57wf\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.869523 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.869630 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.869682 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.894891 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data" (OuterVolumeSpecName: "config-data") pod "d2695413-bac9-4098-8399-72819bda9615" (UID: "d2695413-bac9-4098-8399-72819bda9615"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:09:19 crc kubenswrapper[4960]: I1002 08:09:19.972913 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2695413-bac9-4098-8399-72819bda9615-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.347544 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad1f362d-1d99-4948-a74e-cb656fbc3825" path="/var/lib/kubelet/pods/ad1f362d-1d99-4948-a74e-cb656fbc3825/volumes" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.607146 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.608794 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"29e1c556-f157-4c20-93a7-d0ceecc98439","Type":"ContainerStarted","Data":"c9d953868f3cdfd7c84f52bd466d37bcdc3e42c1a9cbf6f226b8d179fa9d387f"} Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.608940 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"29e1c556-f157-4c20-93a7-d0ceecc98439","Type":"ContainerStarted","Data":"2281dc87a189232349ed02acbf616277636ae7af9a5bd3d80acfeef91c50f9fd"} Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.703272 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.703252751 podStartE2EDuration="2.703252751s" podCreationTimestamp="2025-10-02 08:09:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:09:20.68487642 +0000 UTC m=+3181.716822707" watchObservedRunningTime="2025-10-02 08:09:20.703252751 +0000 UTC m=+3181.735199038" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.715633 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.734942 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.745281 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:09:20 crc kubenswrapper[4960]: E1002 08:09:20.745949 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="probe" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.745969 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="probe" Oct 02 08:09:20 crc kubenswrapper[4960]: E1002 08:09:20.746009 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="manila-scheduler" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.746018 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="manila-scheduler" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.746252 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="probe" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.746276 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2695413-bac9-4098-8399-72819bda9615" containerName="manila-scheduler" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.747767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.755581 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.770415 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.786903 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.787015 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-scripts\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.787040 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c97772e1-2ae5-425b-8532-be4fbc404d15-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.787064 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-config-data\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.787151 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.787219 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhj77\" (UniqueName: \"kubernetes.io/projected/c97772e1-2ae5-425b-8532-be4fbc404d15-kube-api-access-dhj77\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.888561 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.888626 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhj77\" (UniqueName: \"kubernetes.io/projected/c97772e1-2ae5-425b-8532-be4fbc404d15-kube-api-access-dhj77\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.888728 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.888794 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-scripts\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.888819 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c97772e1-2ae5-425b-8532-be4fbc404d15-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.888840 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-config-data\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.888940 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c97772e1-2ae5-425b-8532-be4fbc404d15-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.894893 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-config-data\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.895213 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-scripts\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.896432 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.896743 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c97772e1-2ae5-425b-8532-be4fbc404d15-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:20 crc kubenswrapper[4960]: I1002 08:09:20.906136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhj77\" (UniqueName: \"kubernetes.io/projected/c97772e1-2ae5-425b-8532-be4fbc404d15-kube-api-access-dhj77\") pod \"manila-scheduler-0\" (UID: \"c97772e1-2ae5-425b-8532-be4fbc404d15\") " pod="openstack/manila-scheduler-0" Oct 02 08:09:21 crc kubenswrapper[4960]: I1002 08:09:21.084569 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 08:09:21 crc kubenswrapper[4960]: I1002 08:09:21.391979 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 02 08:09:21 crc kubenswrapper[4960]: I1002 08:09:21.596214 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 08:09:21 crc kubenswrapper[4960]: W1002 08:09:21.602211 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc97772e1_2ae5_425b_8532_be4fbc404d15.slice/crio-7b3aec75c63120927899e255ec6ae4e126a271750e9b6afec7e0f370a22b8dd1 WatchSource:0}: Error finding container 7b3aec75c63120927899e255ec6ae4e126a271750e9b6afec7e0f370a22b8dd1: Status 404 returned error can't find the container with id 7b3aec75c63120927899e255ec6ae4e126a271750e9b6afec7e0f370a22b8dd1 Oct 02 08:09:21 crc kubenswrapper[4960]: I1002 08:09:21.637674 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c97772e1-2ae5-425b-8532-be4fbc404d15","Type":"ContainerStarted","Data":"7b3aec75c63120927899e255ec6ae4e126a271750e9b6afec7e0f370a22b8dd1"} Oct 02 08:09:22 crc kubenswrapper[4960]: I1002 08:09:22.353853 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2695413-bac9-4098-8399-72819bda9615" path="/var/lib/kubelet/pods/d2695413-bac9-4098-8399-72819bda9615/volumes" Oct 02 08:09:22 crc kubenswrapper[4960]: I1002 08:09:22.656788 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c97772e1-2ae5-425b-8532-be4fbc404d15","Type":"ContainerStarted","Data":"bbd9589ab897fc622a9807ff1fb8b258b3b2a5544e27163faec6e1c82e4524e9"} Oct 02 08:09:22 crc kubenswrapper[4960]: I1002 08:09:22.657693 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c97772e1-2ae5-425b-8532-be4fbc404d15","Type":"ContainerStarted","Data":"fe5e1e845532ba63d47026e7d196d773000ee48897e1c4b53560878d0d97ca16"} Oct 02 08:09:22 crc kubenswrapper[4960]: I1002 08:09:22.685458 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.6854272999999997 podStartE2EDuration="2.6854273s" podCreationTimestamp="2025-10-02 08:09:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:09:22.684768892 +0000 UTC m=+3183.716715189" watchObservedRunningTime="2025-10-02 08:09:22.6854273 +0000 UTC m=+3183.717373597" Oct 02 08:09:26 crc kubenswrapper[4960]: E1002 08:09:26.486402 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice/crio-869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:09:28 crc kubenswrapper[4960]: I1002 08:09:28.995185 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 02 08:09:29 crc kubenswrapper[4960]: I1002 08:09:29.330328 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:09:29 crc kubenswrapper[4960]: E1002 08:09:29.330613 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:09:31 crc kubenswrapper[4960]: I1002 08:09:31.085748 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 02 08:09:36 crc kubenswrapper[4960]: E1002 08:09:36.767644 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice/crio-869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8\": RecentStats: unable to find data in memory cache]" Oct 02 08:09:40 crc kubenswrapper[4960]: I1002 08:09:40.578626 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 02 08:09:41 crc kubenswrapper[4960]: I1002 08:09:41.153923 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 08:09:41 crc kubenswrapper[4960]: I1002 08:09:41.330519 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:09:41 crc kubenswrapper[4960]: E1002 08:09:41.330769 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:09:42 crc kubenswrapper[4960]: I1002 08:09:42.678180 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 02 08:09:47 crc kubenswrapper[4960]: E1002 08:09:47.054721 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice/crio-869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:09:54 crc kubenswrapper[4960]: I1002 08:09:54.330297 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:09:54 crc kubenswrapper[4960]: E1002 08:09:54.331509 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:09:57 crc kubenswrapper[4960]: E1002 08:09:57.288868 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice/crio-869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:10:05 crc kubenswrapper[4960]: I1002 08:10:05.330385 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:10:06 crc kubenswrapper[4960]: I1002 08:10:06.162408 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"4caf78e82bca40e6ecc8a0038e596cb9043f062a02c0b5157a1680a789b00dcc"} Oct 02 08:10:07 crc kubenswrapper[4960]: E1002 08:10:07.551500 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice/crio-869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:10:17 crc kubenswrapper[4960]: E1002 08:10:17.812393 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice/crio-869826b8057a8609fa4e5fbe4ef54254f93a2184c66a0c64efa9fc0a885acae8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2695413_bac9_4098_8399_72819bda9615.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:10:43 crc kubenswrapper[4960]: I1002 08:10:43.398960 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz"] Oct 02 08:10:43 crc kubenswrapper[4960]: I1002 08:10:43.402762 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" Oct 02 08:10:43 crc kubenswrapper[4960]: I1002 08:10:43.456291 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz"] Oct 02 08:10:43 crc kubenswrapper[4960]: I1002 08:10:43.599624 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5hvl\" (UniqueName: \"kubernetes.io/projected/cefbfffd-77e3-4fef-a4b6-b4f94a7cee80-kube-api-access-w5hvl\") pod \"openstack-operator-controller-operator-57f4f67bc5-p8hvz\" (UID: \"cefbfffd-77e3-4fef-a4b6-b4f94a7cee80\") " pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" Oct 02 08:10:43 crc kubenswrapper[4960]: I1002 08:10:43.702417 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5hvl\" (UniqueName: \"kubernetes.io/projected/cefbfffd-77e3-4fef-a4b6-b4f94a7cee80-kube-api-access-w5hvl\") pod \"openstack-operator-controller-operator-57f4f67bc5-p8hvz\" (UID: \"cefbfffd-77e3-4fef-a4b6-b4f94a7cee80\") " pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" Oct 02 08:10:43 crc kubenswrapper[4960]: I1002 08:10:43.723312 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5hvl\" (UniqueName: \"kubernetes.io/projected/cefbfffd-77e3-4fef-a4b6-b4f94a7cee80-kube-api-access-w5hvl\") pod \"openstack-operator-controller-operator-57f4f67bc5-p8hvz\" (UID: \"cefbfffd-77e3-4fef-a4b6-b4f94a7cee80\") " pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" Oct 02 08:10:43 crc kubenswrapper[4960]: I1002 08:10:43.736456 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" Oct 02 08:10:44 crc kubenswrapper[4960]: I1002 08:10:44.243199 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz"] Oct 02 08:10:44 crc kubenswrapper[4960]: I1002 08:10:44.603746 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" event={"ID":"cefbfffd-77e3-4fef-a4b6-b4f94a7cee80","Type":"ContainerStarted","Data":"8cab10ae72d97dd506d060340f2fe1fe9b82f648423b685f753f33369597cefc"} Oct 02 08:10:44 crc kubenswrapper[4960]: I1002 08:10:44.604176 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" event={"ID":"cefbfffd-77e3-4fef-a4b6-b4f94a7cee80","Type":"ContainerStarted","Data":"84b6aa268991ad4e46a33d8ae6029c6a5c72a22fa539460606ebd0972be09e47"} Oct 02 08:10:45 crc kubenswrapper[4960]: I1002 08:10:45.618785 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" event={"ID":"cefbfffd-77e3-4fef-a4b6-b4f94a7cee80","Type":"ContainerStarted","Data":"515e81c16c5e18e2dd18fe5e02ad861e679e54d765890e37ed28f6fa512fef51"} Oct 02 08:10:45 crc kubenswrapper[4960]: I1002 08:10:45.619396 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" Oct 02 08:10:45 crc kubenswrapper[4960]: I1002 08:10:45.668719 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" podStartSLOduration=2.668682237 podStartE2EDuration="2.668682237s" podCreationTimestamp="2025-10-02 08:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:10:45.647678555 +0000 UTC m=+3266.679624862" watchObservedRunningTime="2025-10-02 08:10:45.668682237 +0000 UTC m=+3266.700628534" Oct 02 08:10:53 crc kubenswrapper[4960]: I1002 08:10:53.740884 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-57f4f67bc5-p8hvz" Oct 02 08:10:53 crc kubenswrapper[4960]: I1002 08:10:53.838178 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr"] Oct 02 08:10:53 crc kubenswrapper[4960]: I1002 08:10:53.838507 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="operator" containerID="cri-o://25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf" gracePeriod=10 Oct 02 08:10:53 crc kubenswrapper[4960]: I1002 08:10:53.838999 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="kube-rbac-proxy" containerID="cri-o://e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04" gracePeriod=10 Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.378350 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.548501 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrfss\" (UniqueName: \"kubernetes.io/projected/a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54-kube-api-access-wrfss\") pod \"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54\" (UID: \"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54\") " Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.558008 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54-kube-api-access-wrfss" (OuterVolumeSpecName: "kube-api-access-wrfss") pod "a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" (UID: "a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54"). InnerVolumeSpecName "kube-api-access-wrfss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.651034 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrfss\" (UniqueName: \"kubernetes.io/projected/a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54-kube-api-access-wrfss\") on node \"crc\" DevicePath \"\"" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.708043 4960 generic.go:334] "Generic (PLEG): container finished" podID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerID="e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04" exitCode=0 Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.708093 4960 generic.go:334] "Generic (PLEG): container finished" podID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerID="25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf" exitCode=0 Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.708086 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" event={"ID":"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54","Type":"ContainerDied","Data":"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04"} Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.708139 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" event={"ID":"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54","Type":"ContainerDied","Data":"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf"} Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.708156 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" event={"ID":"a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54","Type":"ContainerDied","Data":"1b6b673893251c9c3e6ea00ef5fba1f4363862085dd9ec71ca6c0f7968a61bae"} Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.708131 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.708204 4960 scope.go:117] "RemoveContainer" containerID="e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.741649 4960 scope.go:117] "RemoveContainer" containerID="25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.753078 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr"] Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.766089 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-859455d779-rrmmr"] Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.780594 4960 scope.go:117] "RemoveContainer" containerID="e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04" Oct 02 08:10:54 crc kubenswrapper[4960]: E1002 08:10:54.781955 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04\": container with ID starting with e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04 not found: ID does not exist" containerID="e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.782111 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04"} err="failed to get container status \"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04\": rpc error: code = NotFound desc = could not find container \"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04\": container with ID starting with e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04 not found: ID does not exist" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.782217 4960 scope.go:117] "RemoveContainer" containerID="25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf" Oct 02 08:10:54 crc kubenswrapper[4960]: E1002 08:10:54.782940 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf\": container with ID starting with 25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf not found: ID does not exist" containerID="25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.783137 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf"} err="failed to get container status \"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf\": rpc error: code = NotFound desc = could not find container \"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf\": container with ID starting with 25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf not found: ID does not exist" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.783182 4960 scope.go:117] "RemoveContainer" containerID="e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.783698 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04"} err="failed to get container status \"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04\": rpc error: code = NotFound desc = could not find container \"e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04\": container with ID starting with e2b0bc330d146dd082313694d66ee063e6a2878c5e693904ce79d9c51ff57b04 not found: ID does not exist" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.783777 4960 scope.go:117] "RemoveContainer" containerID="25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf" Oct 02 08:10:54 crc kubenswrapper[4960]: I1002 08:10:54.784393 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf"} err="failed to get container status \"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf\": rpc error: code = NotFound desc = could not find container \"25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf\": container with ID starting with 25fb6ed0bf870c7a0c836ebe34ec84b548d92762ee151a5497b2bd52db93ffbf not found: ID does not exist" Oct 02 08:10:56 crc kubenswrapper[4960]: I1002 08:10:56.340457 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" path="/var/lib/kubelet/pods/a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54/volumes" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.130725 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wvcbp"] Oct 02 08:11:18 crc kubenswrapper[4960]: E1002 08:11:18.131931 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="kube-rbac-proxy" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.131946 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="kube-rbac-proxy" Oct 02 08:11:18 crc kubenswrapper[4960]: E1002 08:11:18.131965 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="operator" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.131993 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="operator" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.132272 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="kube-rbac-proxy" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.132289 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8fcd1e0-0c9e-453f-8e24-cfa92f46fe54" containerName="operator" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.135125 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.159099 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvcbp"] Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.238006 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-catalog-content\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.238080 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-utilities\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.238122 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzfkc\" (UniqueName: \"kubernetes.io/projected/74495065-e84e-4952-8699-492d8552b367-kube-api-access-mzfkc\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.340506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-catalog-content\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.340788 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-utilities\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.340843 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzfkc\" (UniqueName: \"kubernetes.io/projected/74495065-e84e-4952-8699-492d8552b367-kube-api-access-mzfkc\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.341154 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-catalog-content\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.341256 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-utilities\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.371528 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzfkc\" (UniqueName: \"kubernetes.io/projected/74495065-e84e-4952-8699-492d8552b367-kube-api-access-mzfkc\") pod \"redhat-marketplace-wvcbp\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.481429 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.965311 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvcbp"] Oct 02 08:11:18 crc kubenswrapper[4960]: I1002 08:11:18.996364 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvcbp" event={"ID":"74495065-e84e-4952-8699-492d8552b367","Type":"ContainerStarted","Data":"a3e661aac9dde614ae95babe896c0d719eee7fbfc8bc5696cafb0a92c19c39d0"} Oct 02 08:11:20 crc kubenswrapper[4960]: I1002 08:11:20.010299 4960 generic.go:334] "Generic (PLEG): container finished" podID="74495065-e84e-4952-8699-492d8552b367" containerID="68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7" exitCode=0 Oct 02 08:11:20 crc kubenswrapper[4960]: I1002 08:11:20.010913 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvcbp" event={"ID":"74495065-e84e-4952-8699-492d8552b367","Type":"ContainerDied","Data":"68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7"} Oct 02 08:11:21 crc kubenswrapper[4960]: I1002 08:11:21.021526 4960 generic.go:334] "Generic (PLEG): container finished" podID="74495065-e84e-4952-8699-492d8552b367" containerID="47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc" exitCode=0 Oct 02 08:11:21 crc kubenswrapper[4960]: I1002 08:11:21.021580 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvcbp" event={"ID":"74495065-e84e-4952-8699-492d8552b367","Type":"ContainerDied","Data":"47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc"} Oct 02 08:11:22 crc kubenswrapper[4960]: I1002 08:11:22.038157 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvcbp" event={"ID":"74495065-e84e-4952-8699-492d8552b367","Type":"ContainerStarted","Data":"dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10"} Oct 02 08:11:22 crc kubenswrapper[4960]: I1002 08:11:22.065443 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wvcbp" podStartSLOduration=2.569795193 podStartE2EDuration="4.065420558s" podCreationTimestamp="2025-10-02 08:11:18 +0000 UTC" firstStartedPulling="2025-10-02 08:11:20.012919858 +0000 UTC m=+3301.044866165" lastFinishedPulling="2025-10-02 08:11:21.508545233 +0000 UTC m=+3302.540491530" observedRunningTime="2025-10-02 08:11:22.056470689 +0000 UTC m=+3303.088416976" watchObservedRunningTime="2025-10-02 08:11:22.065420558 +0000 UTC m=+3303.097366845" Oct 02 08:11:28 crc kubenswrapper[4960]: I1002 08:11:28.482191 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:28 crc kubenswrapper[4960]: I1002 08:11:28.483475 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:28 crc kubenswrapper[4960]: I1002 08:11:28.540809 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:29 crc kubenswrapper[4960]: I1002 08:11:29.198302 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:29 crc kubenswrapper[4960]: I1002 08:11:29.264489 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvcbp"] Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.140555 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wvcbp" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="registry-server" containerID="cri-o://dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10" gracePeriod=2 Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.634820 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.644438 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9"] Oct 02 08:11:31 crc kubenswrapper[4960]: E1002 08:11:31.644965 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="extract-content" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.645128 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="extract-content" Oct 02 08:11:31 crc kubenswrapper[4960]: E1002 08:11:31.645165 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="extract-utilities" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.645172 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="extract-utilities" Oct 02 08:11:31 crc kubenswrapper[4960]: E1002 08:11:31.645199 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="registry-server" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.645206 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="registry-server" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.645403 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="74495065-e84e-4952-8699-492d8552b367" containerName="registry-server" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.646666 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.667304 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9"] Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.812990 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzfkc\" (UniqueName: \"kubernetes.io/projected/74495065-e84e-4952-8699-492d8552b367-kube-api-access-mzfkc\") pod \"74495065-e84e-4952-8699-492d8552b367\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.813117 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-utilities\") pod \"74495065-e84e-4952-8699-492d8552b367\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.813151 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-catalog-content\") pod \"74495065-e84e-4952-8699-492d8552b367\" (UID: \"74495065-e84e-4952-8699-492d8552b367\") " Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.813599 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rmsh\" (UniqueName: \"kubernetes.io/projected/868e4045-9e39-4add-9023-4ef41caed499-kube-api-access-7rmsh\") pod \"test-operator-controller-manager-fd7b549cd-82rf9\" (UID: \"868e4045-9e39-4add-9023-4ef41caed499\") " pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.813896 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-utilities" (OuterVolumeSpecName: "utilities") pod "74495065-e84e-4952-8699-492d8552b367" (UID: "74495065-e84e-4952-8699-492d8552b367"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.825218 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74495065-e84e-4952-8699-492d8552b367-kube-api-access-mzfkc" (OuterVolumeSpecName: "kube-api-access-mzfkc") pod "74495065-e84e-4952-8699-492d8552b367" (UID: "74495065-e84e-4952-8699-492d8552b367"). InnerVolumeSpecName "kube-api-access-mzfkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.830016 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74495065-e84e-4952-8699-492d8552b367" (UID: "74495065-e84e-4952-8699-492d8552b367"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.915912 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rmsh\" (UniqueName: \"kubernetes.io/projected/868e4045-9e39-4add-9023-4ef41caed499-kube-api-access-7rmsh\") pod \"test-operator-controller-manager-fd7b549cd-82rf9\" (UID: \"868e4045-9e39-4add-9023-4ef41caed499\") " pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.916044 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzfkc\" (UniqueName: \"kubernetes.io/projected/74495065-e84e-4952-8699-492d8552b367-kube-api-access-mzfkc\") on node \"crc\" DevicePath \"\"" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.916057 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.916067 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74495065-e84e-4952-8699-492d8552b367-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.932207 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rmsh\" (UniqueName: \"kubernetes.io/projected/868e4045-9e39-4add-9023-4ef41caed499-kube-api-access-7rmsh\") pod \"test-operator-controller-manager-fd7b549cd-82rf9\" (UID: \"868e4045-9e39-4add-9023-4ef41caed499\") " pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" Oct 02 08:11:31 crc kubenswrapper[4960]: I1002 08:11:31.982955 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.153019 4960 generic.go:334] "Generic (PLEG): container finished" podID="74495065-e84e-4952-8699-492d8552b367" containerID="dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10" exitCode=0 Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.153095 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvcbp" event={"ID":"74495065-e84e-4952-8699-492d8552b367","Type":"ContainerDied","Data":"dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10"} Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.153146 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvcbp" event={"ID":"74495065-e84e-4952-8699-492d8552b367","Type":"ContainerDied","Data":"a3e661aac9dde614ae95babe896c0d719eee7fbfc8bc5696cafb0a92c19c39d0"} Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.153167 4960 scope.go:117] "RemoveContainer" containerID="dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.153288 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvcbp" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.191335 4960 scope.go:117] "RemoveContainer" containerID="47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.197235 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvcbp"] Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.209003 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvcbp"] Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.223351 4960 scope.go:117] "RemoveContainer" containerID="68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.242059 4960 scope.go:117] "RemoveContainer" containerID="dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10" Oct 02 08:11:32 crc kubenswrapper[4960]: E1002 08:11:32.242442 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10\": container with ID starting with dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10 not found: ID does not exist" containerID="dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.242503 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10"} err="failed to get container status \"dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10\": rpc error: code = NotFound desc = could not find container \"dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10\": container with ID starting with dcdb14ddee67d2cddc8b2992846213af1fd763e44f7f64100ce0c7470e55fb10 not found: ID does not exist" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.242537 4960 scope.go:117] "RemoveContainer" containerID="47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc" Oct 02 08:11:32 crc kubenswrapper[4960]: E1002 08:11:32.242815 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc\": container with ID starting with 47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc not found: ID does not exist" containerID="47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.242847 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc"} err="failed to get container status \"47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc\": rpc error: code = NotFound desc = could not find container \"47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc\": container with ID starting with 47a4ccaff6ca3d3de25e53c8c3e65dabbf392074824ab2ef63e6f4e23492edfc not found: ID does not exist" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.242866 4960 scope.go:117] "RemoveContainer" containerID="68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7" Oct 02 08:11:32 crc kubenswrapper[4960]: E1002 08:11:32.243133 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7\": container with ID starting with 68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7 not found: ID does not exist" containerID="68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.243161 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7"} err="failed to get container status \"68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7\": rpc error: code = NotFound desc = could not find container \"68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7\": container with ID starting with 68d80cbe4145680100dce956b96b1a10e37ff6eb4ccaf17cfc975e997cf4b7c7 not found: ID does not exist" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.340647 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74495065-e84e-4952-8699-492d8552b367" path="/var/lib/kubelet/pods/74495065-e84e-4952-8699-492d8552b367/volumes" Oct 02 08:11:32 crc kubenswrapper[4960]: I1002 08:11:32.416966 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9"] Oct 02 08:11:33 crc kubenswrapper[4960]: I1002 08:11:33.162394 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" event={"ID":"868e4045-9e39-4add-9023-4ef41caed499","Type":"ContainerStarted","Data":"db4654b09835724c0608f864975c24a1f76aaddfada02b32ed2d01ea12a3b052"} Oct 02 08:11:34 crc kubenswrapper[4960]: I1002 08:11:34.173828 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" event={"ID":"868e4045-9e39-4add-9023-4ef41caed499","Type":"ContainerStarted","Data":"62a92564aa37adea2f6f5502f5999e2e453fc4c60983a419a8db35fee8afb432"} Oct 02 08:11:34 crc kubenswrapper[4960]: I1002 08:11:34.174142 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" event={"ID":"868e4045-9e39-4add-9023-4ef41caed499","Type":"ContainerStarted","Data":"a8e4cb18c0778bfcec22889ae1cfeca01ed4b1155b5f36b80b9ecbf5577aee94"} Oct 02 08:11:34 crc kubenswrapper[4960]: I1002 08:11:34.174285 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" Oct 02 08:11:34 crc kubenswrapper[4960]: I1002 08:11:34.203761 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" podStartSLOduration=2.201645051 podStartE2EDuration="3.203730001s" podCreationTimestamp="2025-10-02 08:11:31 +0000 UTC" firstStartedPulling="2025-10-02 08:11:32.425411895 +0000 UTC m=+3313.457358182" lastFinishedPulling="2025-10-02 08:11:33.427496835 +0000 UTC m=+3314.459443132" observedRunningTime="2025-10-02 08:11:34.187168268 +0000 UTC m=+3315.219114565" watchObservedRunningTime="2025-10-02 08:11:34.203730001 +0000 UTC m=+3315.235676308" Oct 02 08:11:41 crc kubenswrapper[4960]: I1002 08:11:41.986179 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-fd7b549cd-82rf9" Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.083539 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w"] Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.087732 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="kube-rbac-proxy" containerID="cri-o://4a62e2cb6aa360185567b4af0271b870f66e1ca04f25291798992462c4c8049c" gracePeriod=10 Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.088065 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="manager" containerID="cri-o://35b92ac4bf724f020ead22c4ceda7e25d661fbd0079bb6e60143c399455bd99b" gracePeriod=10 Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.310931 4960 generic.go:334] "Generic (PLEG): container finished" podID="5529d258-1683-402a-ba4d-fa355405201d" containerID="35b92ac4bf724f020ead22c4ceda7e25d661fbd0079bb6e60143c399455bd99b" exitCode=0 Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.310960 4960 generic.go:334] "Generic (PLEG): container finished" podID="5529d258-1683-402a-ba4d-fa355405201d" containerID="4a62e2cb6aa360185567b4af0271b870f66e1ca04f25291798992462c4c8049c" exitCode=0 Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.311014 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" event={"ID":"5529d258-1683-402a-ba4d-fa355405201d","Type":"ContainerDied","Data":"35b92ac4bf724f020ead22c4ceda7e25d661fbd0079bb6e60143c399455bd99b"} Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.311044 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" event={"ID":"5529d258-1683-402a-ba4d-fa355405201d","Type":"ContainerDied","Data":"4a62e2cb6aa360185567b4af0271b870f66e1ca04f25291798992462c4c8049c"} Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.544188 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.673573 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l7sf\" (UniqueName: \"kubernetes.io/projected/5529d258-1683-402a-ba4d-fa355405201d-kube-api-access-4l7sf\") pod \"5529d258-1683-402a-ba4d-fa355405201d\" (UID: \"5529d258-1683-402a-ba4d-fa355405201d\") " Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.681919 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5529d258-1683-402a-ba4d-fa355405201d-kube-api-access-4l7sf" (OuterVolumeSpecName: "kube-api-access-4l7sf") pod "5529d258-1683-402a-ba4d-fa355405201d" (UID: "5529d258-1683-402a-ba4d-fa355405201d"). InnerVolumeSpecName "kube-api-access-4l7sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:11:42 crc kubenswrapper[4960]: I1002 08:11:42.776589 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l7sf\" (UniqueName: \"kubernetes.io/projected/5529d258-1683-402a-ba4d-fa355405201d-kube-api-access-4l7sf\") on node \"crc\" DevicePath \"\"" Oct 02 08:11:43 crc kubenswrapper[4960]: I1002 08:11:43.325891 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" event={"ID":"5529d258-1683-402a-ba4d-fa355405201d","Type":"ContainerDied","Data":"4e5ed95462f180a6a82e3a5f0bac04ac8794bba44ab007a0eb96865d3e54f770"} Oct 02 08:11:43 crc kubenswrapper[4960]: I1002 08:11:43.325961 4960 scope.go:117] "RemoveContainer" containerID="35b92ac4bf724f020ead22c4ceda7e25d661fbd0079bb6e60143c399455bd99b" Oct 02 08:11:43 crc kubenswrapper[4960]: I1002 08:11:43.326072 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w" Oct 02 08:11:43 crc kubenswrapper[4960]: I1002 08:11:43.377902 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w"] Oct 02 08:11:43 crc kubenswrapper[4960]: I1002 08:11:43.381522 4960 scope.go:117] "RemoveContainer" containerID="4a62e2cb6aa360185567b4af0271b870f66e1ca04f25291798992462c4c8049c" Oct 02 08:11:43 crc kubenswrapper[4960]: I1002 08:11:43.389401 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-sqb6w"] Oct 02 08:11:44 crc kubenswrapper[4960]: I1002 08:11:44.389899 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5529d258-1683-402a-ba4d-fa355405201d" path="/var/lib/kubelet/pods/5529d258-1683-402a-ba4d-fa355405201d/volumes" Oct 02 08:12:29 crc kubenswrapper[4960]: I1002 08:12:29.150484 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:12:29 crc kubenswrapper[4960]: I1002 08:12:29.152423 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.422465 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hvxj7"] Oct 02 08:12:48 crc kubenswrapper[4960]: E1002 08:12:48.423415 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="kube-rbac-proxy" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.423428 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="kube-rbac-proxy" Oct 02 08:12:48 crc kubenswrapper[4960]: E1002 08:12:48.423472 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="manager" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.423478 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="manager" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.423649 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="manager" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.423666 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5529d258-1683-402a-ba4d-fa355405201d" containerName="kube-rbac-proxy" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.425023 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.433907 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hvxj7"] Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.541888 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-utilities\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.542047 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-catalog-content\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.542117 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfsqw\" (UniqueName: \"kubernetes.io/projected/e09acca7-db30-40af-91ff-a5becac7023f-kube-api-access-xfsqw\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.643818 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-utilities\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.644209 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-catalog-content\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.644252 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfsqw\" (UniqueName: \"kubernetes.io/projected/e09acca7-db30-40af-91ff-a5becac7023f-kube-api-access-xfsqw\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.644487 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-utilities\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.644923 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-catalog-content\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.673861 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfsqw\" (UniqueName: \"kubernetes.io/projected/e09acca7-db30-40af-91ff-a5becac7023f-kube-api-access-xfsqw\") pod \"certified-operators-hvxj7\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:48 crc kubenswrapper[4960]: I1002 08:12:48.760412 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:49 crc kubenswrapper[4960]: I1002 08:12:49.282115 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hvxj7"] Oct 02 08:12:50 crc kubenswrapper[4960]: I1002 08:12:50.056839 4960 generic.go:334] "Generic (PLEG): container finished" podID="e09acca7-db30-40af-91ff-a5becac7023f" containerID="e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5" exitCode=0 Oct 02 08:12:50 crc kubenswrapper[4960]: I1002 08:12:50.056950 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvxj7" event={"ID":"e09acca7-db30-40af-91ff-a5becac7023f","Type":"ContainerDied","Data":"e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5"} Oct 02 08:12:50 crc kubenswrapper[4960]: I1002 08:12:50.057283 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvxj7" event={"ID":"e09acca7-db30-40af-91ff-a5becac7023f","Type":"ContainerStarted","Data":"3ff5f554b31dff809cf3e5fb3a0f093194af22f7ce73ba214c58ab299f76c45a"} Oct 02 08:12:51 crc kubenswrapper[4960]: I1002 08:12:51.075910 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvxj7" event={"ID":"e09acca7-db30-40af-91ff-a5becac7023f","Type":"ContainerStarted","Data":"c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5"} Oct 02 08:12:52 crc kubenswrapper[4960]: I1002 08:12:52.092216 4960 generic.go:334] "Generic (PLEG): container finished" podID="e09acca7-db30-40af-91ff-a5becac7023f" containerID="c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5" exitCode=0 Oct 02 08:12:52 crc kubenswrapper[4960]: I1002 08:12:52.092638 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvxj7" event={"ID":"e09acca7-db30-40af-91ff-a5becac7023f","Type":"ContainerDied","Data":"c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5"} Oct 02 08:12:53 crc kubenswrapper[4960]: I1002 08:12:53.116223 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvxj7" event={"ID":"e09acca7-db30-40af-91ff-a5becac7023f","Type":"ContainerStarted","Data":"4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc"} Oct 02 08:12:53 crc kubenswrapper[4960]: I1002 08:12:53.166645 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hvxj7" podStartSLOduration=2.496429044 podStartE2EDuration="5.166603167s" podCreationTimestamp="2025-10-02 08:12:48 +0000 UTC" firstStartedPulling="2025-10-02 08:12:50.059333848 +0000 UTC m=+3391.091280165" lastFinishedPulling="2025-10-02 08:12:52.729508001 +0000 UTC m=+3393.761454288" observedRunningTime="2025-10-02 08:12:53.137132657 +0000 UTC m=+3394.169078964" watchObservedRunningTime="2025-10-02 08:12:53.166603167 +0000 UTC m=+3394.198549494" Oct 02 08:12:58 crc kubenswrapper[4960]: I1002 08:12:58.762300 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:58 crc kubenswrapper[4960]: I1002 08:12:58.763158 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:58 crc kubenswrapper[4960]: I1002 08:12:58.834896 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:59 crc kubenswrapper[4960]: I1002 08:12:59.150081 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:12:59 crc kubenswrapper[4960]: I1002 08:12:59.150445 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:12:59 crc kubenswrapper[4960]: I1002 08:12:59.254120 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:12:59 crc kubenswrapper[4960]: I1002 08:12:59.319363 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hvxj7"] Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.194029 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hvxj7" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="registry-server" containerID="cri-o://4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc" gracePeriod=2 Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.755799 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.860896 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-utilities\") pod \"e09acca7-db30-40af-91ff-a5becac7023f\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.861084 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfsqw\" (UniqueName: \"kubernetes.io/projected/e09acca7-db30-40af-91ff-a5becac7023f-kube-api-access-xfsqw\") pod \"e09acca7-db30-40af-91ff-a5becac7023f\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.861210 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-catalog-content\") pod \"e09acca7-db30-40af-91ff-a5becac7023f\" (UID: \"e09acca7-db30-40af-91ff-a5becac7023f\") " Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.861890 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-utilities" (OuterVolumeSpecName: "utilities") pod "e09acca7-db30-40af-91ff-a5becac7023f" (UID: "e09acca7-db30-40af-91ff-a5becac7023f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.873476 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e09acca7-db30-40af-91ff-a5becac7023f-kube-api-access-xfsqw" (OuterVolumeSpecName: "kube-api-access-xfsqw") pod "e09acca7-db30-40af-91ff-a5becac7023f" (UID: "e09acca7-db30-40af-91ff-a5becac7023f"). InnerVolumeSpecName "kube-api-access-xfsqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.914592 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e09acca7-db30-40af-91ff-a5becac7023f" (UID: "e09acca7-db30-40af-91ff-a5becac7023f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.962938 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfsqw\" (UniqueName: \"kubernetes.io/projected/e09acca7-db30-40af-91ff-a5becac7023f-kube-api-access-xfsqw\") on node \"crc\" DevicePath \"\"" Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.962987 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:13:01 crc kubenswrapper[4960]: I1002 08:13:01.962997 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e09acca7-db30-40af-91ff-a5becac7023f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.202867 4960 generic.go:334] "Generic (PLEG): container finished" podID="e09acca7-db30-40af-91ff-a5becac7023f" containerID="4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc" exitCode=0 Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.202912 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvxj7" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.202911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvxj7" event={"ID":"e09acca7-db30-40af-91ff-a5becac7023f","Type":"ContainerDied","Data":"4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc"} Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.203023 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvxj7" event={"ID":"e09acca7-db30-40af-91ff-a5becac7023f","Type":"ContainerDied","Data":"3ff5f554b31dff809cf3e5fb3a0f093194af22f7ce73ba214c58ab299f76c45a"} Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.203040 4960 scope.go:117] "RemoveContainer" containerID="4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.235849 4960 scope.go:117] "RemoveContainer" containerID="c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.238915 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hvxj7"] Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.251494 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hvxj7"] Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.259869 4960 scope.go:117] "RemoveContainer" containerID="e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.304588 4960 scope.go:117] "RemoveContainer" containerID="4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc" Oct 02 08:13:02 crc kubenswrapper[4960]: E1002 08:13:02.305106 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc\": container with ID starting with 4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc not found: ID does not exist" containerID="4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.305156 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc"} err="failed to get container status \"4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc\": rpc error: code = NotFound desc = could not find container \"4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc\": container with ID starting with 4fd66ed2000d7245157d9fbc56d49fadffa7a9dfa1a61b4fc1461e067009bbcc not found: ID does not exist" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.305182 4960 scope.go:117] "RemoveContainer" containerID="c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5" Oct 02 08:13:02 crc kubenswrapper[4960]: E1002 08:13:02.305560 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5\": container with ID starting with c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5 not found: ID does not exist" containerID="c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.305581 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5"} err="failed to get container status \"c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5\": rpc error: code = NotFound desc = could not find container \"c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5\": container with ID starting with c7fb2ba4fa59f33f8ba7bfbf3d7cad610d199efcbb914a952483ecc8274d5ae5 not found: ID does not exist" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.305612 4960 scope.go:117] "RemoveContainer" containerID="e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5" Oct 02 08:13:02 crc kubenswrapper[4960]: E1002 08:13:02.305917 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5\": container with ID starting with e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5 not found: ID does not exist" containerID="e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.305937 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5"} err="failed to get container status \"e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5\": rpc error: code = NotFound desc = could not find container \"e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5\": container with ID starting with e5841d8a90e121de49dcdbf4f4df7be297229dce5dba9ee3b276e6322e7243f5 not found: ID does not exist" Oct 02 08:13:02 crc kubenswrapper[4960]: I1002 08:13:02.356479 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e09acca7-db30-40af-91ff-a5becac7023f" path="/var/lib/kubelet/pods/e09acca7-db30-40af-91ff-a5becac7023f/volumes" Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.150299 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.151280 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.151406 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.152862 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4caf78e82bca40e6ecc8a0038e596cb9043f062a02c0b5157a1680a789b00dcc"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.153026 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://4caf78e82bca40e6ecc8a0038e596cb9043f062a02c0b5157a1680a789b00dcc" gracePeriod=600 Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.518888 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="4caf78e82bca40e6ecc8a0038e596cb9043f062a02c0b5157a1680a789b00dcc" exitCode=0 Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.519008 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"4caf78e82bca40e6ecc8a0038e596cb9043f062a02c0b5157a1680a789b00dcc"} Oct 02 08:13:29 crc kubenswrapper[4960]: I1002 08:13:29.519085 4960 scope.go:117] "RemoveContainer" containerID="3a613fad859a5e975b9d609728284ce27c1caaeb7cecd6089e7380d265491498" Oct 02 08:13:30 crc kubenswrapper[4960]: I1002 08:13:30.535029 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9"} Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.932069 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 02 08:13:34 crc kubenswrapper[4960]: E1002 08:13:34.933138 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="registry-server" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.933154 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="registry-server" Oct 02 08:13:34 crc kubenswrapper[4960]: E1002 08:13:34.933190 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="extract-content" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.933198 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="extract-content" Oct 02 08:13:34 crc kubenswrapper[4960]: E1002 08:13:34.933219 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="extract-utilities" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.933227 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="extract-utilities" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.933452 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09acca7-db30-40af-91ff-a5becac7023f" containerName="registry-server" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.934194 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.937318 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.937335 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vk2jb" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.945084 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.945106 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 08:13:34 crc kubenswrapper[4960]: I1002 08:13:34.956594 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.012998 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013337 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013421 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013462 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013500 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013540 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013586 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013646 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-428rb\" (UniqueName: \"kubernetes.io/projected/aed3697d-fab0-4c3e-b5d0-21fb0037a412-kube-api-access-428rb\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013668 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.013739 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.115894 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.116357 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.116576 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.116736 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.117005 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.117200 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.117402 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.117646 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.117781 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.118104 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.118131 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.117077 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.118370 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-428rb\" (UniqueName: \"kubernetes.io/projected/aed3697d-fab0-4c3e-b5d0-21fb0037a412-kube-api-access-428rb\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.118602 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.119915 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.125461 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.125546 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.128418 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.129455 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.147086 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-428rb\" (UniqueName: \"kubernetes.io/projected/aed3697d-fab0-4c3e-b5d0-21fb0037a412-kube-api-access-428rb\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.156014 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.265363 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.886218 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 02 08:13:35 crc kubenswrapper[4960]: I1002 08:13:35.893447 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:13:36 crc kubenswrapper[4960]: I1002 08:13:36.597183 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"aed3697d-fab0-4c3e-b5d0-21fb0037a412","Type":"ContainerStarted","Data":"e6ed1391854bf9612376cd0f4ac9102c01b00222994ebdfee93b4f0d6788212c"} Oct 02 08:14:02 crc kubenswrapper[4960]: E1002 08:14:02.130052 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 02 08:14:02 crc kubenswrapper[4960]: E1002 08:14:02.132350 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-428rb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-full_openstack(aed3697d-fab0-4c3e-b5d0-21fb0037a412): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 08:14:02 crc kubenswrapper[4960]: E1002 08:14:02.134120 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="aed3697d-fab0-4c3e-b5d0-21fb0037a412" Oct 02 08:14:02 crc kubenswrapper[4960]: E1002 08:14:02.864140 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="aed3697d-fab0-4c3e-b5d0-21fb0037a412" Oct 02 08:14:20 crc kubenswrapper[4960]: I1002 08:14:20.051129 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"aed3697d-fab0-4c3e-b5d0-21fb0037a412","Type":"ContainerStarted","Data":"eedcd7a8b6bba3120d5e7cc924e5e82f258f2541e8e637a43675452ca77dc609"} Oct 02 08:14:20 crc kubenswrapper[4960]: I1002 08:14:20.085032 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-full" podStartSLOduration=4.165865827 podStartE2EDuration="47.085000555s" podCreationTimestamp="2025-10-02 08:13:33 +0000 UTC" firstStartedPulling="2025-10-02 08:13:35.893145823 +0000 UTC m=+3436.925092130" lastFinishedPulling="2025-10-02 08:14:18.812280541 +0000 UTC m=+3479.844226858" observedRunningTime="2025-10-02 08:14:20.075048639 +0000 UTC m=+3481.106994946" watchObservedRunningTime="2025-10-02 08:14:20.085000555 +0000 UTC m=+3481.116946842" Oct 02 08:14:22 crc kubenswrapper[4960]: I1002 08:14:22.914055 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tlcgh"] Oct 02 08:14:22 crc kubenswrapper[4960]: I1002 08:14:22.918811 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:22 crc kubenswrapper[4960]: I1002 08:14:22.936058 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tlcgh"] Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.013390 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-catalog-content\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.013475 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njktv\" (UniqueName: \"kubernetes.io/projected/791d95b4-0f54-4f45-bd4a-740993156764-kube-api-access-njktv\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.013574 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-utilities\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.114942 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-catalog-content\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.115055 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njktv\" (UniqueName: \"kubernetes.io/projected/791d95b4-0f54-4f45-bd4a-740993156764-kube-api-access-njktv\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.115141 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-utilities\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.115619 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-catalog-content\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.115696 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-utilities\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.141468 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njktv\" (UniqueName: \"kubernetes.io/projected/791d95b4-0f54-4f45-bd4a-740993156764-kube-api-access-njktv\") pod \"redhat-operators-tlcgh\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.267112 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:23 crc kubenswrapper[4960]: I1002 08:14:23.782741 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tlcgh"] Oct 02 08:14:24 crc kubenswrapper[4960]: I1002 08:14:24.108316 4960 generic.go:334] "Generic (PLEG): container finished" podID="791d95b4-0f54-4f45-bd4a-740993156764" containerID="2d987174dbf86f1535481719f808be7cd7f4aa959355cb1effe84300ab601d12" exitCode=0 Oct 02 08:14:24 crc kubenswrapper[4960]: I1002 08:14:24.108366 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlcgh" event={"ID":"791d95b4-0f54-4f45-bd4a-740993156764","Type":"ContainerDied","Data":"2d987174dbf86f1535481719f808be7cd7f4aa959355cb1effe84300ab601d12"} Oct 02 08:14:24 crc kubenswrapper[4960]: I1002 08:14:24.108398 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlcgh" event={"ID":"791d95b4-0f54-4f45-bd4a-740993156764","Type":"ContainerStarted","Data":"37af6eddb3982ca0735d0ac75d1a307c9d34913926283c6889d646875ad5df99"} Oct 02 08:14:26 crc kubenswrapper[4960]: I1002 08:14:26.137202 4960 generic.go:334] "Generic (PLEG): container finished" podID="791d95b4-0f54-4f45-bd4a-740993156764" containerID="2c788fa95b8150b01a0abe148c7e7f6aa7b859b2594b22f17adf6bbb9bc96655" exitCode=0 Oct 02 08:14:26 crc kubenswrapper[4960]: I1002 08:14:26.137303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlcgh" event={"ID":"791d95b4-0f54-4f45-bd4a-740993156764","Type":"ContainerDied","Data":"2c788fa95b8150b01a0abe148c7e7f6aa7b859b2594b22f17adf6bbb9bc96655"} Oct 02 08:14:27 crc kubenswrapper[4960]: I1002 08:14:27.152211 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlcgh" event={"ID":"791d95b4-0f54-4f45-bd4a-740993156764","Type":"ContainerStarted","Data":"351e738cccbdb4a2afff19aaa1207dcfc327427e5cf7b6778f22706c4268bd5a"} Oct 02 08:14:27 crc kubenswrapper[4960]: I1002 08:14:27.187743 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tlcgh" podStartSLOduration=2.7863414410000003 podStartE2EDuration="5.187716509s" podCreationTimestamp="2025-10-02 08:14:22 +0000 UTC" firstStartedPulling="2025-10-02 08:14:24.111339098 +0000 UTC m=+3485.143285385" lastFinishedPulling="2025-10-02 08:14:26.512714136 +0000 UTC m=+3487.544660453" observedRunningTime="2025-10-02 08:14:27.179447027 +0000 UTC m=+3488.211393304" watchObservedRunningTime="2025-10-02 08:14:27.187716509 +0000 UTC m=+3488.219662826" Oct 02 08:14:33 crc kubenswrapper[4960]: I1002 08:14:33.268363 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:33 crc kubenswrapper[4960]: I1002 08:14:33.269266 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:33 crc kubenswrapper[4960]: I1002 08:14:33.346041 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:34 crc kubenswrapper[4960]: I1002 08:14:34.308327 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:34 crc kubenswrapper[4960]: I1002 08:14:34.388908 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tlcgh"] Oct 02 08:14:36 crc kubenswrapper[4960]: I1002 08:14:36.267405 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tlcgh" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="registry-server" containerID="cri-o://351e738cccbdb4a2afff19aaa1207dcfc327427e5cf7b6778f22706c4268bd5a" gracePeriod=2 Oct 02 08:14:37 crc kubenswrapper[4960]: I1002 08:14:37.278119 4960 generic.go:334] "Generic (PLEG): container finished" podID="791d95b4-0f54-4f45-bd4a-740993156764" containerID="351e738cccbdb4a2afff19aaa1207dcfc327427e5cf7b6778f22706c4268bd5a" exitCode=0 Oct 02 08:14:37 crc kubenswrapper[4960]: I1002 08:14:37.278235 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlcgh" event={"ID":"791d95b4-0f54-4f45-bd4a-740993156764","Type":"ContainerDied","Data":"351e738cccbdb4a2afff19aaa1207dcfc327427e5cf7b6778f22706c4268bd5a"} Oct 02 08:14:37 crc kubenswrapper[4960]: I1002 08:14:37.910831 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.035902 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-utilities\") pod \"791d95b4-0f54-4f45-bd4a-740993156764\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.036003 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-catalog-content\") pod \"791d95b4-0f54-4f45-bd4a-740993156764\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.036311 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njktv\" (UniqueName: \"kubernetes.io/projected/791d95b4-0f54-4f45-bd4a-740993156764-kube-api-access-njktv\") pod \"791d95b4-0f54-4f45-bd4a-740993156764\" (UID: \"791d95b4-0f54-4f45-bd4a-740993156764\") " Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.037696 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-utilities" (OuterVolumeSpecName: "utilities") pod "791d95b4-0f54-4f45-bd4a-740993156764" (UID: "791d95b4-0f54-4f45-bd4a-740993156764"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.045375 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791d95b4-0f54-4f45-bd4a-740993156764-kube-api-access-njktv" (OuterVolumeSpecName: "kube-api-access-njktv") pod "791d95b4-0f54-4f45-bd4a-740993156764" (UID: "791d95b4-0f54-4f45-bd4a-740993156764"). InnerVolumeSpecName "kube-api-access-njktv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.133722 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "791d95b4-0f54-4f45-bd4a-740993156764" (UID: "791d95b4-0f54-4f45-bd4a-740993156764"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.138136 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.138179 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791d95b4-0f54-4f45-bd4a-740993156764-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.138191 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njktv\" (UniqueName: \"kubernetes.io/projected/791d95b4-0f54-4f45-bd4a-740993156764-kube-api-access-njktv\") on node \"crc\" DevicePath \"\"" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.300260 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlcgh" event={"ID":"791d95b4-0f54-4f45-bd4a-740993156764","Type":"ContainerDied","Data":"37af6eddb3982ca0735d0ac75d1a307c9d34913926283c6889d646875ad5df99"} Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.300344 4960 scope.go:117] "RemoveContainer" containerID="351e738cccbdb4a2afff19aaa1207dcfc327427e5cf7b6778f22706c4268bd5a" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.300389 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlcgh" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.353223 4960 scope.go:117] "RemoveContainer" containerID="2c788fa95b8150b01a0abe148c7e7f6aa7b859b2594b22f17adf6bbb9bc96655" Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.356139 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tlcgh"] Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.357675 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tlcgh"] Oct 02 08:14:38 crc kubenswrapper[4960]: I1002 08:14:38.378854 4960 scope.go:117] "RemoveContainer" containerID="2d987174dbf86f1535481719f808be7cd7f4aa959355cb1effe84300ab601d12" Oct 02 08:14:40 crc kubenswrapper[4960]: I1002 08:14:40.343620 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="791d95b4-0f54-4f45-bd4a-740993156764" path="/var/lib/kubelet/pods/791d95b4-0f54-4f45-bd4a-740993156764/volumes" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.179726 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh"] Oct 02 08:15:00 crc kubenswrapper[4960]: E1002 08:15:00.180821 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="extract-content" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.180837 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="extract-content" Oct 02 08:15:00 crc kubenswrapper[4960]: E1002 08:15:00.180854 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="registry-server" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.180862 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="registry-server" Oct 02 08:15:00 crc kubenswrapper[4960]: E1002 08:15:00.180884 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="extract-utilities" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.180893 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="extract-utilities" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.181212 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="791d95b4-0f54-4f45-bd4a-740993156764" containerName="registry-server" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.182014 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.185455 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.185548 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.190482 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh"] Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.326445 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-secret-volume\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.326570 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-config-volume\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.326598 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwff5\" (UniqueName: \"kubernetes.io/projected/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-kube-api-access-fwff5\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.428864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-secret-volume\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.428990 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-config-volume\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.429010 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwff5\" (UniqueName: \"kubernetes.io/projected/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-kube-api-access-fwff5\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.429817 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-config-volume\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.438404 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-secret-volume\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.444884 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwff5\" (UniqueName: \"kubernetes.io/projected/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-kube-api-access-fwff5\") pod \"collect-profiles-29323215-j25kh\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.509397 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.613856 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n24sb"] Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.616385 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.629063 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n24sb"] Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.741124 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrjjb\" (UniqueName: \"kubernetes.io/projected/7d77fff6-b3ba-461a-8000-85c34162c857-kube-api-access-xrjjb\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.741206 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-utilities\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.741303 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-catalog-content\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.843417 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrjjb\" (UniqueName: \"kubernetes.io/projected/7d77fff6-b3ba-461a-8000-85c34162c857-kube-api-access-xrjjb\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.843486 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-utilities\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.843570 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-catalog-content\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.844009 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-catalog-content\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.844125 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-utilities\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.862440 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrjjb\" (UniqueName: \"kubernetes.io/projected/7d77fff6-b3ba-461a-8000-85c34162c857-kube-api-access-xrjjb\") pod \"community-operators-n24sb\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.969416 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:00 crc kubenswrapper[4960]: I1002 08:15:00.999338 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh"] Oct 02 08:15:01 crc kubenswrapper[4960]: I1002 08:15:01.440443 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n24sb"] Oct 02 08:15:01 crc kubenswrapper[4960]: W1002 08:15:01.450051 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d77fff6_b3ba_461a_8000_85c34162c857.slice/crio-a66b3eb83440ea0d37c6c94a0655ec6128f0c2aeed2207c6493485b956f70b56 WatchSource:0}: Error finding container a66b3eb83440ea0d37c6c94a0655ec6128f0c2aeed2207c6493485b956f70b56: Status 404 returned error can't find the container with id a66b3eb83440ea0d37c6c94a0655ec6128f0c2aeed2207c6493485b956f70b56 Oct 02 08:15:01 crc kubenswrapper[4960]: I1002 08:15:01.536474 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n24sb" event={"ID":"7d77fff6-b3ba-461a-8000-85c34162c857","Type":"ContainerStarted","Data":"a66b3eb83440ea0d37c6c94a0655ec6128f0c2aeed2207c6493485b956f70b56"} Oct 02 08:15:01 crc kubenswrapper[4960]: I1002 08:15:01.538434 4960 generic.go:334] "Generic (PLEG): container finished" podID="e85da6fb-ac9e-49f3-813c-2f249b9e44ec" containerID="3477d3b78361f5d88ce543d6318ad3c220a6ed682bbe7e1e9c09e022d00cefdf" exitCode=0 Oct 02 08:15:01 crc kubenswrapper[4960]: I1002 08:15:01.538482 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" event={"ID":"e85da6fb-ac9e-49f3-813c-2f249b9e44ec","Type":"ContainerDied","Data":"3477d3b78361f5d88ce543d6318ad3c220a6ed682bbe7e1e9c09e022d00cefdf"} Oct 02 08:15:01 crc kubenswrapper[4960]: I1002 08:15:01.538527 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" event={"ID":"e85da6fb-ac9e-49f3-813c-2f249b9e44ec","Type":"ContainerStarted","Data":"217ff9d2505c1d561f19a1c201ef687d00b6fe52abaa139259879bc0c8733cf9"} Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.108529 4960 scope.go:117] "RemoveContainer" containerID="30ad267bc1c8360e2537d61a27384fe5b9144dbb9eb8adcef8afd8711a6752de" Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.130102 4960 scope.go:117] "RemoveContainer" containerID="083cc1cd63078041bcb54566f3f1f0878218809774cf45d62c9d1792520b4aa3" Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.550573 4960 generic.go:334] "Generic (PLEG): container finished" podID="7d77fff6-b3ba-461a-8000-85c34162c857" containerID="5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23" exitCode=0 Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.550784 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n24sb" event={"ID":"7d77fff6-b3ba-461a-8000-85c34162c857","Type":"ContainerDied","Data":"5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23"} Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.881134 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.988626 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-config-volume\") pod \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.988807 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwff5\" (UniqueName: \"kubernetes.io/projected/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-kube-api-access-fwff5\") pod \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.988877 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-secret-volume\") pod \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\" (UID: \"e85da6fb-ac9e-49f3-813c-2f249b9e44ec\") " Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.990207 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-config-volume" (OuterVolumeSpecName: "config-volume") pod "e85da6fb-ac9e-49f3-813c-2f249b9e44ec" (UID: "e85da6fb-ac9e-49f3-813c-2f249b9e44ec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.994075 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e85da6fb-ac9e-49f3-813c-2f249b9e44ec" (UID: "e85da6fb-ac9e-49f3-813c-2f249b9e44ec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:15:02 crc kubenswrapper[4960]: I1002 08:15:02.999191 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-kube-api-access-fwff5" (OuterVolumeSpecName: "kube-api-access-fwff5") pod "e85da6fb-ac9e-49f3-813c-2f249b9e44ec" (UID: "e85da6fb-ac9e-49f3-813c-2f249b9e44ec"). InnerVolumeSpecName "kube-api-access-fwff5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.091705 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.091737 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwff5\" (UniqueName: \"kubernetes.io/projected/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-kube-api-access-fwff5\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.091749 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e85da6fb-ac9e-49f3-813c-2f249b9e44ec-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.561251 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" event={"ID":"e85da6fb-ac9e-49f3-813c-2f249b9e44ec","Type":"ContainerDied","Data":"217ff9d2505c1d561f19a1c201ef687d00b6fe52abaa139259879bc0c8733cf9"} Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.561603 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="217ff9d2505c1d561f19a1c201ef687d00b6fe52abaa139259879bc0c8733cf9" Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.561451 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh" Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.948760 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k"] Oct 02 08:15:03 crc kubenswrapper[4960]: I1002 08:15:03.957056 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-dzb9k"] Oct 02 08:15:04 crc kubenswrapper[4960]: I1002 08:15:04.343156 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b65ad47-92d7-42f5-97d7-dc83810d1bc6" path="/var/lib/kubelet/pods/2b65ad47-92d7-42f5-97d7-dc83810d1bc6/volumes" Oct 02 08:15:04 crc kubenswrapper[4960]: I1002 08:15:04.571084 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n24sb" event={"ID":"7d77fff6-b3ba-461a-8000-85c34162c857","Type":"ContainerStarted","Data":"9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315"} Oct 02 08:15:05 crc kubenswrapper[4960]: I1002 08:15:05.585484 4960 generic.go:334] "Generic (PLEG): container finished" podID="7d77fff6-b3ba-461a-8000-85c34162c857" containerID="9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315" exitCode=0 Oct 02 08:15:05 crc kubenswrapper[4960]: I1002 08:15:05.585574 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n24sb" event={"ID":"7d77fff6-b3ba-461a-8000-85c34162c857","Type":"ContainerDied","Data":"9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315"} Oct 02 08:15:06 crc kubenswrapper[4960]: I1002 08:15:06.596042 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n24sb" event={"ID":"7d77fff6-b3ba-461a-8000-85c34162c857","Type":"ContainerStarted","Data":"1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28"} Oct 02 08:15:06 crc kubenswrapper[4960]: I1002 08:15:06.614015 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n24sb" podStartSLOduration=3.163393221 podStartE2EDuration="6.61395754s" podCreationTimestamp="2025-10-02 08:15:00 +0000 UTC" firstStartedPulling="2025-10-02 08:15:02.55264743 +0000 UTC m=+3523.584593727" lastFinishedPulling="2025-10-02 08:15:06.003211759 +0000 UTC m=+3527.035158046" observedRunningTime="2025-10-02 08:15:06.612456089 +0000 UTC m=+3527.644402396" watchObservedRunningTime="2025-10-02 08:15:06.61395754 +0000 UTC m=+3527.645903847" Oct 02 08:15:10 crc kubenswrapper[4960]: I1002 08:15:10.970334 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:10 crc kubenswrapper[4960]: I1002 08:15:10.970813 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:11 crc kubenswrapper[4960]: I1002 08:15:11.016739 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:11 crc kubenswrapper[4960]: I1002 08:15:11.705398 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:11 crc kubenswrapper[4960]: I1002 08:15:11.760620 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n24sb"] Oct 02 08:15:13 crc kubenswrapper[4960]: I1002 08:15:13.677451 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n24sb" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="registry-server" containerID="cri-o://1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28" gracePeriod=2 Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.192562 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.245208 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-utilities\") pod \"7d77fff6-b3ba-461a-8000-85c34162c857\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.245406 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrjjb\" (UniqueName: \"kubernetes.io/projected/7d77fff6-b3ba-461a-8000-85c34162c857-kube-api-access-xrjjb\") pod \"7d77fff6-b3ba-461a-8000-85c34162c857\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.245452 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-catalog-content\") pod \"7d77fff6-b3ba-461a-8000-85c34162c857\" (UID: \"7d77fff6-b3ba-461a-8000-85c34162c857\") " Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.247951 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-utilities" (OuterVolumeSpecName: "utilities") pod "7d77fff6-b3ba-461a-8000-85c34162c857" (UID: "7d77fff6-b3ba-461a-8000-85c34162c857"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.254497 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d77fff6-b3ba-461a-8000-85c34162c857-kube-api-access-xrjjb" (OuterVolumeSpecName: "kube-api-access-xrjjb") pod "7d77fff6-b3ba-461a-8000-85c34162c857" (UID: "7d77fff6-b3ba-461a-8000-85c34162c857"). InnerVolumeSpecName "kube-api-access-xrjjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.348250 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrjjb\" (UniqueName: \"kubernetes.io/projected/7d77fff6-b3ba-461a-8000-85c34162c857-kube-api-access-xrjjb\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.348276 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.371685 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d77fff6-b3ba-461a-8000-85c34162c857" (UID: "7d77fff6-b3ba-461a-8000-85c34162c857"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.450024 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d77fff6-b3ba-461a-8000-85c34162c857-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.686422 4960 generic.go:334] "Generic (PLEG): container finished" podID="7d77fff6-b3ba-461a-8000-85c34162c857" containerID="1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28" exitCode=0 Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.686467 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n24sb" event={"ID":"7d77fff6-b3ba-461a-8000-85c34162c857","Type":"ContainerDied","Data":"1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28"} Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.686500 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n24sb" event={"ID":"7d77fff6-b3ba-461a-8000-85c34162c857","Type":"ContainerDied","Data":"a66b3eb83440ea0d37c6c94a0655ec6128f0c2aeed2207c6493485b956f70b56"} Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.686522 4960 scope.go:117] "RemoveContainer" containerID="1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.686559 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n24sb" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.718037 4960 scope.go:117] "RemoveContainer" containerID="9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.727132 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n24sb"] Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.735566 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n24sb"] Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.755294 4960 scope.go:117] "RemoveContainer" containerID="5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.781759 4960 scope.go:117] "RemoveContainer" containerID="1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28" Oct 02 08:15:14 crc kubenswrapper[4960]: E1002 08:15:14.782236 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28\": container with ID starting with 1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28 not found: ID does not exist" containerID="1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.782306 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28"} err="failed to get container status \"1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28\": rpc error: code = NotFound desc = could not find container \"1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28\": container with ID starting with 1ecedd7613ba0e288b41c523481618418cf83a8f1a4a2b64c5871b1f7e749f28 not found: ID does not exist" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.782348 4960 scope.go:117] "RemoveContainer" containerID="9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315" Oct 02 08:15:14 crc kubenswrapper[4960]: E1002 08:15:14.782786 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315\": container with ID starting with 9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315 not found: ID does not exist" containerID="9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.782807 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315"} err="failed to get container status \"9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315\": rpc error: code = NotFound desc = could not find container \"9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315\": container with ID starting with 9dd675dd75f8f077ee379ff2bc20ccedb75f9c4d0e68df664ea57836e64a0315 not found: ID does not exist" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.782854 4960 scope.go:117] "RemoveContainer" containerID="5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23" Oct 02 08:15:14 crc kubenswrapper[4960]: E1002 08:15:14.783260 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23\": container with ID starting with 5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23 not found: ID does not exist" containerID="5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23" Oct 02 08:15:14 crc kubenswrapper[4960]: I1002 08:15:14.783283 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23"} err="failed to get container status \"5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23\": rpc error: code = NotFound desc = could not find container \"5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23\": container with ID starting with 5242b2bfa2dfaea1e249e2c92a042168e95bb74364f1a2d05b8bfa8ed3679b23 not found: ID does not exist" Oct 02 08:15:16 crc kubenswrapper[4960]: I1002 08:15:16.338785 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" path="/var/lib/kubelet/pods/7d77fff6-b3ba-461a-8000-85c34162c857/volumes" Oct 02 08:15:29 crc kubenswrapper[4960]: I1002 08:15:29.150314 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:15:29 crc kubenswrapper[4960]: I1002 08:15:29.151028 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:15:59 crc kubenswrapper[4960]: I1002 08:15:59.150566 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:15:59 crc kubenswrapper[4960]: I1002 08:15:59.151223 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:16:02 crc kubenswrapper[4960]: I1002 08:16:02.187715 4960 scope.go:117] "RemoveContainer" containerID="ba2cfc4f3f33a09417513f041a1d9069be25c54b809775aa0436ebc4a1d2f566" Oct 02 08:16:02 crc kubenswrapper[4960]: I1002 08:16:02.217780 4960 scope.go:117] "RemoveContainer" containerID="c89050772bc026f2534ff2a2ea5f3b2b9bfe53273cb1474a5081a7fd5d10adbb" Oct 02 08:16:02 crc kubenswrapper[4960]: I1002 08:16:02.283004 4960 scope.go:117] "RemoveContainer" containerID="a9ed8b134a23a0ccef28e6585652bec0710917d13c7bce87b2409049c33594db" Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.150544 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.151024 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.151069 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.151663 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.151719 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" gracePeriod=600 Oct 02 08:16:29 crc kubenswrapper[4960]: E1002 08:16:29.284780 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.435301 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" exitCode=0 Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.435347 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9"} Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.435382 4960 scope.go:117] "RemoveContainer" containerID="4caf78e82bca40e6ecc8a0038e596cb9043f062a02c0b5157a1680a789b00dcc" Oct 02 08:16:29 crc kubenswrapper[4960]: I1002 08:16:29.436094 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:16:29 crc kubenswrapper[4960]: E1002 08:16:29.436386 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:16:44 crc kubenswrapper[4960]: I1002 08:16:44.330894 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:16:44 crc kubenswrapper[4960]: E1002 08:16:44.331815 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:16:56 crc kubenswrapper[4960]: I1002 08:16:56.329434 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:16:56 crc kubenswrapper[4960]: E1002 08:16:56.330204 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:17:11 crc kubenswrapper[4960]: I1002 08:17:11.330503 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:17:11 crc kubenswrapper[4960]: E1002 08:17:11.331853 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:17:25 crc kubenswrapper[4960]: I1002 08:17:25.329917 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:17:25 crc kubenswrapper[4960]: E1002 08:17:25.330832 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:17:37 crc kubenswrapper[4960]: I1002 08:17:37.330103 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:17:37 crc kubenswrapper[4960]: E1002 08:17:37.331240 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:17:52 crc kubenswrapper[4960]: I1002 08:17:52.329805 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:17:52 crc kubenswrapper[4960]: E1002 08:17:52.330742 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:18:06 crc kubenswrapper[4960]: I1002 08:18:06.330729 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:18:06 crc kubenswrapper[4960]: E1002 08:18:06.331678 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:18:20 crc kubenswrapper[4960]: I1002 08:18:20.343268 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:18:20 crc kubenswrapper[4960]: E1002 08:18:20.344130 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:18:21 crc kubenswrapper[4960]: I1002 08:18:21.044359 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-7h55n"] Oct 02 08:18:21 crc kubenswrapper[4960]: I1002 08:18:21.056046 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-7h55n"] Oct 02 08:18:22 crc kubenswrapper[4960]: I1002 08:18:22.352102 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f90db36-1706-4086-8289-91cdf46731ac" path="/var/lib/kubelet/pods/0f90db36-1706-4086-8289-91cdf46731ac/volumes" Oct 02 08:18:33 crc kubenswrapper[4960]: I1002 08:18:33.028165 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-8be1-account-create-gpbfg"] Oct 02 08:18:33 crc kubenswrapper[4960]: I1002 08:18:33.042314 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-8be1-account-create-gpbfg"] Oct 02 08:18:33 crc kubenswrapper[4960]: I1002 08:18:33.330670 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:18:33 crc kubenswrapper[4960]: E1002 08:18:33.331080 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:18:34 crc kubenswrapper[4960]: I1002 08:18:34.349889 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc4a490-fbec-448d-bbcf-e436060d9497" path="/var/lib/kubelet/pods/3fc4a490-fbec-448d-bbcf-e436060d9497/volumes" Oct 02 08:18:46 crc kubenswrapper[4960]: I1002 08:18:46.330632 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:18:46 crc kubenswrapper[4960]: E1002 08:18:46.331413 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:18:53 crc kubenswrapper[4960]: I1002 08:18:53.045315 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-czszs"] Oct 02 08:18:53 crc kubenswrapper[4960]: I1002 08:18:53.055645 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-czszs"] Oct 02 08:18:54 crc kubenswrapper[4960]: I1002 08:18:54.348065 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72581d48-c6bf-4140-a8d1-11c058d33638" path="/var/lib/kubelet/pods/72581d48-c6bf-4140-a8d1-11c058d33638/volumes" Oct 02 08:18:59 crc kubenswrapper[4960]: I1002 08:18:59.331590 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:18:59 crc kubenswrapper[4960]: E1002 08:18:59.332754 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:19:02 crc kubenswrapper[4960]: I1002 08:19:02.425057 4960 scope.go:117] "RemoveContainer" containerID="204a8ce30c11934c4044fe935f2d320cf579434ab72b53513c4abb86de0ea028" Oct 02 08:19:02 crc kubenswrapper[4960]: I1002 08:19:02.455387 4960 scope.go:117] "RemoveContainer" containerID="b9dff3d366277357551605ff08101eb77268824acc29c1945e0284d2d04a367d" Oct 02 08:19:02 crc kubenswrapper[4960]: I1002 08:19:02.513328 4960 scope.go:117] "RemoveContainer" containerID="f07f2966e81ba9ed92661bf3a39a7bd1d865abae4a9a0f05916d398658c8a6ac" Oct 02 08:19:14 crc kubenswrapper[4960]: I1002 08:19:14.336141 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:19:14 crc kubenswrapper[4960]: E1002 08:19:14.337404 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:19:25 crc kubenswrapper[4960]: I1002 08:19:25.330110 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:19:25 crc kubenswrapper[4960]: E1002 08:19:25.331064 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:19:37 crc kubenswrapper[4960]: I1002 08:19:37.330407 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:19:37 crc kubenswrapper[4960]: E1002 08:19:37.331918 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:19:49 crc kubenswrapper[4960]: I1002 08:19:49.330499 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:19:49 crc kubenswrapper[4960]: E1002 08:19:49.331837 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:20:04 crc kubenswrapper[4960]: I1002 08:20:04.330281 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:20:04 crc kubenswrapper[4960]: E1002 08:20:04.331261 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:20:15 crc kubenswrapper[4960]: I1002 08:20:15.330499 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:20:15 crc kubenswrapper[4960]: E1002 08:20:15.331218 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:20:30 crc kubenswrapper[4960]: I1002 08:20:30.335826 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:20:30 crc kubenswrapper[4960]: E1002 08:20:30.336743 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:20:43 crc kubenswrapper[4960]: I1002 08:20:43.334366 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:20:43 crc kubenswrapper[4960]: E1002 08:20:43.335294 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:20:58 crc kubenswrapper[4960]: I1002 08:20:58.330047 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:20:58 crc kubenswrapper[4960]: E1002 08:20:58.330847 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:21:12 crc kubenswrapper[4960]: I1002 08:21:12.330046 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:21:12 crc kubenswrapper[4960]: E1002 08:21:12.330901 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.193675 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k8cdx"] Oct 02 08:21:21 crc kubenswrapper[4960]: E1002 08:21:21.204728 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e85da6fb-ac9e-49f3-813c-2f249b9e44ec" containerName="collect-profiles" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.204751 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e85da6fb-ac9e-49f3-813c-2f249b9e44ec" containerName="collect-profiles" Oct 02 08:21:21 crc kubenswrapper[4960]: E1002 08:21:21.204769 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="extract-utilities" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.204776 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="extract-utilities" Oct 02 08:21:21 crc kubenswrapper[4960]: E1002 08:21:21.204803 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="registry-server" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.204811 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="registry-server" Oct 02 08:21:21 crc kubenswrapper[4960]: E1002 08:21:21.204826 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="extract-content" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.204833 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="extract-content" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.205070 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e85da6fb-ac9e-49f3-813c-2f249b9e44ec" containerName="collect-profiles" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.205095 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d77fff6-b3ba-461a-8000-85c34162c857" containerName="registry-server" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.206826 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8cdx"] Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.206934 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.287479 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w76lg\" (UniqueName: \"kubernetes.io/projected/68cf3a90-1425-4eea-914a-446182b32cdd-kube-api-access-w76lg\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.287560 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-utilities\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.287923 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-catalog-content\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.389885 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w76lg\" (UniqueName: \"kubernetes.io/projected/68cf3a90-1425-4eea-914a-446182b32cdd-kube-api-access-w76lg\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.389975 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-utilities\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.390097 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-catalog-content\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.390475 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-utilities\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.390529 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-catalog-content\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.413254 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w76lg\" (UniqueName: \"kubernetes.io/projected/68cf3a90-1425-4eea-914a-446182b32cdd-kube-api-access-w76lg\") pod \"redhat-marketplace-k8cdx\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:21 crc kubenswrapper[4960]: I1002 08:21:21.529861 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:22 crc kubenswrapper[4960]: I1002 08:21:22.044239 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8cdx"] Oct 02 08:21:22 crc kubenswrapper[4960]: I1002 08:21:22.264560 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8cdx" event={"ID":"68cf3a90-1425-4eea-914a-446182b32cdd","Type":"ContainerStarted","Data":"cc4096fc84a4dbdc75ef4ca89af8e16513c3ae5631674d325dbd971253d9c706"} Oct 02 08:21:23 crc kubenswrapper[4960]: I1002 08:21:23.278499 4960 generic.go:334] "Generic (PLEG): container finished" podID="68cf3a90-1425-4eea-914a-446182b32cdd" containerID="67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580" exitCode=0 Oct 02 08:21:23 crc kubenswrapper[4960]: I1002 08:21:23.278619 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8cdx" event={"ID":"68cf3a90-1425-4eea-914a-446182b32cdd","Type":"ContainerDied","Data":"67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580"} Oct 02 08:21:23 crc kubenswrapper[4960]: I1002 08:21:23.281820 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:21:25 crc kubenswrapper[4960]: I1002 08:21:25.302556 4960 generic.go:334] "Generic (PLEG): container finished" podID="68cf3a90-1425-4eea-914a-446182b32cdd" containerID="cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb" exitCode=0 Oct 02 08:21:25 crc kubenswrapper[4960]: I1002 08:21:25.302627 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8cdx" event={"ID":"68cf3a90-1425-4eea-914a-446182b32cdd","Type":"ContainerDied","Data":"cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb"} Oct 02 08:21:26 crc kubenswrapper[4960]: I1002 08:21:26.317059 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8cdx" event={"ID":"68cf3a90-1425-4eea-914a-446182b32cdd","Type":"ContainerStarted","Data":"1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780"} Oct 02 08:21:26 crc kubenswrapper[4960]: I1002 08:21:26.330820 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:21:26 crc kubenswrapper[4960]: E1002 08:21:26.331361 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:21:26 crc kubenswrapper[4960]: I1002 08:21:26.354215 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k8cdx" podStartSLOduration=2.96320827 podStartE2EDuration="5.354189848s" podCreationTimestamp="2025-10-02 08:21:21 +0000 UTC" firstStartedPulling="2025-10-02 08:21:23.281354492 +0000 UTC m=+3904.313300799" lastFinishedPulling="2025-10-02 08:21:25.67233609 +0000 UTC m=+3906.704282377" observedRunningTime="2025-10-02 08:21:26.346521622 +0000 UTC m=+3907.378467919" watchObservedRunningTime="2025-10-02 08:21:26.354189848 +0000 UTC m=+3907.386136135" Oct 02 08:21:31 crc kubenswrapper[4960]: I1002 08:21:31.531385 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:31 crc kubenswrapper[4960]: I1002 08:21:31.533070 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:31 crc kubenswrapper[4960]: I1002 08:21:31.582860 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:32 crc kubenswrapper[4960]: I1002 08:21:32.416452 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:32 crc kubenswrapper[4960]: I1002 08:21:32.473793 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8cdx"] Oct 02 08:21:34 crc kubenswrapper[4960]: I1002 08:21:34.389285 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k8cdx" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="registry-server" containerID="cri-o://1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780" gracePeriod=2 Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.160431 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.204114 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w76lg\" (UniqueName: \"kubernetes.io/projected/68cf3a90-1425-4eea-914a-446182b32cdd-kube-api-access-w76lg\") pod \"68cf3a90-1425-4eea-914a-446182b32cdd\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.204362 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-utilities\") pod \"68cf3a90-1425-4eea-914a-446182b32cdd\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.204470 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-catalog-content\") pod \"68cf3a90-1425-4eea-914a-446182b32cdd\" (UID: \"68cf3a90-1425-4eea-914a-446182b32cdd\") " Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.206166 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-utilities" (OuterVolumeSpecName: "utilities") pod "68cf3a90-1425-4eea-914a-446182b32cdd" (UID: "68cf3a90-1425-4eea-914a-446182b32cdd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.235252 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68cf3a90-1425-4eea-914a-446182b32cdd-kube-api-access-w76lg" (OuterVolumeSpecName: "kube-api-access-w76lg") pod "68cf3a90-1425-4eea-914a-446182b32cdd" (UID: "68cf3a90-1425-4eea-914a-446182b32cdd"). InnerVolumeSpecName "kube-api-access-w76lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.282155 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68cf3a90-1425-4eea-914a-446182b32cdd" (UID: "68cf3a90-1425-4eea-914a-446182b32cdd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.306456 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.306505 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w76lg\" (UniqueName: \"kubernetes.io/projected/68cf3a90-1425-4eea-914a-446182b32cdd-kube-api-access-w76lg\") on node \"crc\" DevicePath \"\"" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.306522 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68cf3a90-1425-4eea-914a-446182b32cdd-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.417306 4960 generic.go:334] "Generic (PLEG): container finished" podID="68cf3a90-1425-4eea-914a-446182b32cdd" containerID="1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780" exitCode=0 Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.417355 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8cdx" event={"ID":"68cf3a90-1425-4eea-914a-446182b32cdd","Type":"ContainerDied","Data":"1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780"} Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.417388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8cdx" event={"ID":"68cf3a90-1425-4eea-914a-446182b32cdd","Type":"ContainerDied","Data":"cc4096fc84a4dbdc75ef4ca89af8e16513c3ae5631674d325dbd971253d9c706"} Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.417397 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8cdx" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.417406 4960 scope.go:117] "RemoveContainer" containerID="1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.454818 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8cdx"] Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.467798 4960 scope.go:117] "RemoveContainer" containerID="cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.469476 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8cdx"] Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.495751 4960 scope.go:117] "RemoveContainer" containerID="67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.545183 4960 scope.go:117] "RemoveContainer" containerID="1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780" Oct 02 08:21:35 crc kubenswrapper[4960]: E1002 08:21:35.546761 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780\": container with ID starting with 1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780 not found: ID does not exist" containerID="1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.546806 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780"} err="failed to get container status \"1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780\": rpc error: code = NotFound desc = could not find container \"1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780\": container with ID starting with 1f5b80c1ff016238d45a78c5b9dc57234adcfbc913d033a31bebb9daa8e30780 not found: ID does not exist" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.546838 4960 scope.go:117] "RemoveContainer" containerID="cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb" Oct 02 08:21:35 crc kubenswrapper[4960]: E1002 08:21:35.547407 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb\": container with ID starting with cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb not found: ID does not exist" containerID="cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.547437 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb"} err="failed to get container status \"cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb\": rpc error: code = NotFound desc = could not find container \"cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb\": container with ID starting with cb0933a31a0ebe107914da572ee488249fdcd07be516ed504e130635d60a23bb not found: ID does not exist" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.547470 4960 scope.go:117] "RemoveContainer" containerID="67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580" Oct 02 08:21:35 crc kubenswrapper[4960]: E1002 08:21:35.548012 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580\": container with ID starting with 67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580 not found: ID does not exist" containerID="67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580" Oct 02 08:21:35 crc kubenswrapper[4960]: I1002 08:21:35.548050 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580"} err="failed to get container status \"67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580\": rpc error: code = NotFound desc = could not find container \"67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580\": container with ID starting with 67fcf81fc922b4a143315552732f4afebad805bf84295605db9b29fd103e2580 not found: ID does not exist" Oct 02 08:21:36 crc kubenswrapper[4960]: I1002 08:21:36.340655 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" path="/var/lib/kubelet/pods/68cf3a90-1425-4eea-914a-446182b32cdd/volumes" Oct 02 08:21:40 crc kubenswrapper[4960]: I1002 08:21:40.339322 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:21:41 crc kubenswrapper[4960]: I1002 08:21:41.494512 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"e78d5afcfa8229fda3c1a41a2550cb021603e250d345d7dbd2122e8b207bc886"} Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.561457 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d6d42"] Oct 02 08:23:47 crc kubenswrapper[4960]: E1002 08:23:47.562647 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="extract-utilities" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.562666 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="extract-utilities" Oct 02 08:23:47 crc kubenswrapper[4960]: E1002 08:23:47.562689 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="extract-content" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.562697 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="extract-content" Oct 02 08:23:47 crc kubenswrapper[4960]: E1002 08:23:47.562710 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="registry-server" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.562717 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="registry-server" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.562935 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cf3a90-1425-4eea-914a-446182b32cdd" containerName="registry-server" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.564541 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.587099 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d6d42"] Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.628460 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx65v\" (UniqueName: \"kubernetes.io/projected/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-kube-api-access-mx65v\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.628556 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-catalog-content\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.628713 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-utilities\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.730615 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-utilities\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.730787 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx65v\" (UniqueName: \"kubernetes.io/projected/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-kube-api-access-mx65v\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.730815 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-catalog-content\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.731335 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-utilities\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.731360 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-catalog-content\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.749348 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx65v\" (UniqueName: \"kubernetes.io/projected/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-kube-api-access-mx65v\") pod \"certified-operators-d6d42\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:47 crc kubenswrapper[4960]: I1002 08:23:47.892382 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:48 crc kubenswrapper[4960]: I1002 08:23:48.387250 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d6d42"] Oct 02 08:23:48 crc kubenswrapper[4960]: I1002 08:23:48.727401 4960 generic.go:334] "Generic (PLEG): container finished" podID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerID="3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d" exitCode=0 Oct 02 08:23:48 crc kubenswrapper[4960]: I1002 08:23:48.727486 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6d42" event={"ID":"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12","Type":"ContainerDied","Data":"3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d"} Oct 02 08:23:48 crc kubenswrapper[4960]: I1002 08:23:48.727882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6d42" event={"ID":"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12","Type":"ContainerStarted","Data":"f0b8905e79078b0d2ef86d0e026a092f6994bb3f5c577f174cd53564bf94631a"} Oct 02 08:23:49 crc kubenswrapper[4960]: I1002 08:23:49.738928 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6d42" event={"ID":"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12","Type":"ContainerStarted","Data":"7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9"} Oct 02 08:23:51 crc kubenswrapper[4960]: I1002 08:23:51.759321 4960 generic.go:334] "Generic (PLEG): container finished" podID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerID="7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9" exitCode=0 Oct 02 08:23:51 crc kubenswrapper[4960]: I1002 08:23:51.759521 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6d42" event={"ID":"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12","Type":"ContainerDied","Data":"7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9"} Oct 02 08:23:52 crc kubenswrapper[4960]: I1002 08:23:52.773028 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6d42" event={"ID":"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12","Type":"ContainerStarted","Data":"dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036"} Oct 02 08:23:52 crc kubenswrapper[4960]: I1002 08:23:52.801404 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d6d42" podStartSLOduration=2.3171419970000002 podStartE2EDuration="5.801388163s" podCreationTimestamp="2025-10-02 08:23:47 +0000 UTC" firstStartedPulling="2025-10-02 08:23:48.728878544 +0000 UTC m=+4049.760824851" lastFinishedPulling="2025-10-02 08:23:52.21312473 +0000 UTC m=+4053.245071017" observedRunningTime="2025-10-02 08:23:52.799219724 +0000 UTC m=+4053.831166011" watchObservedRunningTime="2025-10-02 08:23:52.801388163 +0000 UTC m=+4053.833334450" Oct 02 08:23:57 crc kubenswrapper[4960]: I1002 08:23:57.892684 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:57 crc kubenswrapper[4960]: I1002 08:23:57.894565 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:58 crc kubenswrapper[4960]: I1002 08:23:58.051778 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:58 crc kubenswrapper[4960]: I1002 08:23:58.900058 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:23:58 crc kubenswrapper[4960]: I1002 08:23:58.967146 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d6d42"] Oct 02 08:23:59 crc kubenswrapper[4960]: I1002 08:23:59.150939 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:23:59 crc kubenswrapper[4960]: I1002 08:23:59.151015 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:24:00 crc kubenswrapper[4960]: I1002 08:24:00.853807 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d6d42" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="registry-server" containerID="cri-o://dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036" gracePeriod=2 Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.566658 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.691741 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-utilities\") pod \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.692265 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-catalog-content\") pod \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.695111 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx65v\" (UniqueName: \"kubernetes.io/projected/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-kube-api-access-mx65v\") pod \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\" (UID: \"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12\") " Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.693254 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-utilities" (OuterVolumeSpecName: "utilities") pod "6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" (UID: "6cf4fb96-8396-4e2b-854d-5bb6f34ecb12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.698481 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.703367 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-kube-api-access-mx65v" (OuterVolumeSpecName: "kube-api-access-mx65v") pod "6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" (UID: "6cf4fb96-8396-4e2b-854d-5bb6f34ecb12"). InnerVolumeSpecName "kube-api-access-mx65v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.739935 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" (UID: "6cf4fb96-8396-4e2b-854d-5bb6f34ecb12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.800182 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.800222 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx65v\" (UniqueName: \"kubernetes.io/projected/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12-kube-api-access-mx65v\") on node \"crc\" DevicePath \"\"" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.865458 4960 generic.go:334] "Generic (PLEG): container finished" podID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerID="dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036" exitCode=0 Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.865505 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6d42" event={"ID":"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12","Type":"ContainerDied","Data":"dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036"} Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.865536 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6d42" event={"ID":"6cf4fb96-8396-4e2b-854d-5bb6f34ecb12","Type":"ContainerDied","Data":"f0b8905e79078b0d2ef86d0e026a092f6994bb3f5c577f174cd53564bf94631a"} Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.865553 4960 scope.go:117] "RemoveContainer" containerID="dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.865690 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6d42" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.903871 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d6d42"] Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.905425 4960 scope.go:117] "RemoveContainer" containerID="7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.914722 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d6d42"] Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.931421 4960 scope.go:117] "RemoveContainer" containerID="3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.979998 4960 scope.go:117] "RemoveContainer" containerID="dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036" Oct 02 08:24:01 crc kubenswrapper[4960]: E1002 08:24:01.980387 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036\": container with ID starting with dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036 not found: ID does not exist" containerID="dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.980432 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036"} err="failed to get container status \"dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036\": rpc error: code = NotFound desc = could not find container \"dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036\": container with ID starting with dacd8452fd9ffbc3d76f445a9a2584b969339b18588a06ffe0dea70d23b9c036 not found: ID does not exist" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.980461 4960 scope.go:117] "RemoveContainer" containerID="7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9" Oct 02 08:24:01 crc kubenswrapper[4960]: E1002 08:24:01.980768 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9\": container with ID starting with 7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9 not found: ID does not exist" containerID="7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.980905 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9"} err="failed to get container status \"7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9\": rpc error: code = NotFound desc = could not find container \"7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9\": container with ID starting with 7d8d34bba9a61229fce367a2c378053f9ee92006c3b1c7b0fade96d4b820e1a9 not found: ID does not exist" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.981007 4960 scope.go:117] "RemoveContainer" containerID="3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d" Oct 02 08:24:01 crc kubenswrapper[4960]: E1002 08:24:01.981261 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d\": container with ID starting with 3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d not found: ID does not exist" containerID="3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d" Oct 02 08:24:01 crc kubenswrapper[4960]: I1002 08:24:01.981281 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d"} err="failed to get container status \"3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d\": rpc error: code = NotFound desc = could not find container \"3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d\": container with ID starting with 3cdea248b41b70fcc9dd9b1098dfc6cc06ac62194755b8f5b7885077c0d9304d not found: ID does not exist" Oct 02 08:24:02 crc kubenswrapper[4960]: I1002 08:24:02.341148 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" path="/var/lib/kubelet/pods/6cf4fb96-8396-4e2b-854d-5bb6f34ecb12/volumes" Oct 02 08:24:29 crc kubenswrapper[4960]: I1002 08:24:29.150199 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:24:29 crc kubenswrapper[4960]: I1002 08:24:29.151234 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.558237 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-757kx"] Oct 02 08:24:37 crc kubenswrapper[4960]: E1002 08:24:37.559870 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="extract-utilities" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.559888 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="extract-utilities" Oct 02 08:24:37 crc kubenswrapper[4960]: E1002 08:24:37.559913 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="extract-content" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.559921 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="extract-content" Oct 02 08:24:37 crc kubenswrapper[4960]: E1002 08:24:37.559941 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="registry-server" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.559951 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="registry-server" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.560222 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cf4fb96-8396-4e2b-854d-5bb6f34ecb12" containerName="registry-server" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.562231 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.568818 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-757kx"] Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.663870 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp8wq\" (UniqueName: \"kubernetes.io/projected/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-kube-api-access-sp8wq\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.664269 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-catalog-content\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.664640 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-utilities\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.766161 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-catalog-content\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.766492 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-utilities\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.766649 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp8wq\" (UniqueName: \"kubernetes.io/projected/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-kube-api-access-sp8wq\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.766939 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-catalog-content\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.766963 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-utilities\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.794101 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp8wq\" (UniqueName: \"kubernetes.io/projected/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-kube-api-access-sp8wq\") pod \"redhat-operators-757kx\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:37 crc kubenswrapper[4960]: I1002 08:24:37.923779 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:38 crc kubenswrapper[4960]: I1002 08:24:38.437948 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-757kx"] Oct 02 08:24:39 crc kubenswrapper[4960]: I1002 08:24:39.244072 4960 generic.go:334] "Generic (PLEG): container finished" podID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerID="6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b" exitCode=0 Oct 02 08:24:39 crc kubenswrapper[4960]: I1002 08:24:39.244247 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-757kx" event={"ID":"c0b733a5-a8da-4e7e-a08e-18412f72cc1b","Type":"ContainerDied","Data":"6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b"} Oct 02 08:24:39 crc kubenswrapper[4960]: I1002 08:24:39.244505 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-757kx" event={"ID":"c0b733a5-a8da-4e7e-a08e-18412f72cc1b","Type":"ContainerStarted","Data":"b6d3adcff8728377c69195ba766d5fbbf688ff16fe6624c0a7c3f2c6a5520eb7"} Oct 02 08:24:40 crc kubenswrapper[4960]: I1002 08:24:40.257829 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-757kx" event={"ID":"c0b733a5-a8da-4e7e-a08e-18412f72cc1b","Type":"ContainerStarted","Data":"142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098"} Oct 02 08:24:42 crc kubenswrapper[4960]: I1002 08:24:42.274019 4960 generic.go:334] "Generic (PLEG): container finished" podID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerID="142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098" exitCode=0 Oct 02 08:24:42 crc kubenswrapper[4960]: I1002 08:24:42.274312 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-757kx" event={"ID":"c0b733a5-a8da-4e7e-a08e-18412f72cc1b","Type":"ContainerDied","Data":"142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098"} Oct 02 08:24:43 crc kubenswrapper[4960]: I1002 08:24:43.288711 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-757kx" event={"ID":"c0b733a5-a8da-4e7e-a08e-18412f72cc1b","Type":"ContainerStarted","Data":"b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b"} Oct 02 08:24:43 crc kubenswrapper[4960]: I1002 08:24:43.318123 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-757kx" podStartSLOduration=2.866405982 podStartE2EDuration="6.318096826s" podCreationTimestamp="2025-10-02 08:24:37 +0000 UTC" firstStartedPulling="2025-10-02 08:24:39.246740948 +0000 UTC m=+4100.278687235" lastFinishedPulling="2025-10-02 08:24:42.698431802 +0000 UTC m=+4103.730378079" observedRunningTime="2025-10-02 08:24:43.308347055 +0000 UTC m=+4104.340293352" watchObservedRunningTime="2025-10-02 08:24:43.318096826 +0000 UTC m=+4104.350043113" Oct 02 08:24:47 crc kubenswrapper[4960]: I1002 08:24:47.924704 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:47 crc kubenswrapper[4960]: I1002 08:24:47.925782 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:24:48 crc kubenswrapper[4960]: I1002 08:24:48.978285 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-757kx" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="registry-server" probeResult="failure" output=< Oct 02 08:24:48 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 08:24:48 crc kubenswrapper[4960]: > Oct 02 08:24:58 crc kubenswrapper[4960]: I1002 08:24:58.986467 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-757kx" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="registry-server" probeResult="failure" output=< Oct 02 08:24:58 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 08:24:58 crc kubenswrapper[4960]: > Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.150406 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.150477 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.150529 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.151354 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e78d5afcfa8229fda3c1a41a2550cb021603e250d345d7dbd2122e8b207bc886"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.151414 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://e78d5afcfa8229fda3c1a41a2550cb021603e250d345d7dbd2122e8b207bc886" gracePeriod=600 Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.483794 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="e78d5afcfa8229fda3c1a41a2550cb021603e250d345d7dbd2122e8b207bc886" exitCode=0 Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.483894 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"e78d5afcfa8229fda3c1a41a2550cb021603e250d345d7dbd2122e8b207bc886"} Oct 02 08:24:59 crc kubenswrapper[4960]: I1002 08:24:59.484014 4960 scope.go:117] "RemoveContainer" containerID="9c0d82486fec9815523a11deeaef9477eab3059f4e2f7b504c762c2f851e8ae9" Oct 02 08:25:00 crc kubenswrapper[4960]: I1002 08:25:00.498700 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056"} Oct 02 08:25:07 crc kubenswrapper[4960]: I1002 08:25:07.975136 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:25:08 crc kubenswrapper[4960]: I1002 08:25:08.042067 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:25:08 crc kubenswrapper[4960]: I1002 08:25:08.767410 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-757kx"] Oct 02 08:25:09 crc kubenswrapper[4960]: I1002 08:25:09.580235 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-757kx" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="registry-server" containerID="cri-o://b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b" gracePeriod=2 Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.255834 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.373748 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-catalog-content\") pod \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.373913 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp8wq\" (UniqueName: \"kubernetes.io/projected/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-kube-api-access-sp8wq\") pod \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.374108 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-utilities\") pod \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\" (UID: \"c0b733a5-a8da-4e7e-a08e-18412f72cc1b\") " Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.375554 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-utilities" (OuterVolumeSpecName: "utilities") pod "c0b733a5-a8da-4e7e-a08e-18412f72cc1b" (UID: "c0b733a5-a8da-4e7e-a08e-18412f72cc1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.401929 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-kube-api-access-sp8wq" (OuterVolumeSpecName: "kube-api-access-sp8wq") pod "c0b733a5-a8da-4e7e-a08e-18412f72cc1b" (UID: "c0b733a5-a8da-4e7e-a08e-18412f72cc1b"). InnerVolumeSpecName "kube-api-access-sp8wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.449881 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0b733a5-a8da-4e7e-a08e-18412f72cc1b" (UID: "c0b733a5-a8da-4e7e-a08e-18412f72cc1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.477390 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.477429 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp8wq\" (UniqueName: \"kubernetes.io/projected/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-kube-api-access-sp8wq\") on node \"crc\" DevicePath \"\"" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.477442 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0b733a5-a8da-4e7e-a08e-18412f72cc1b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.591361 4960 generic.go:334] "Generic (PLEG): container finished" podID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerID="b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b" exitCode=0 Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.591421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-757kx" event={"ID":"c0b733a5-a8da-4e7e-a08e-18412f72cc1b","Type":"ContainerDied","Data":"b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b"} Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.591459 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-757kx" event={"ID":"c0b733a5-a8da-4e7e-a08e-18412f72cc1b","Type":"ContainerDied","Data":"b6d3adcff8728377c69195ba766d5fbbf688ff16fe6624c0a7c3f2c6a5520eb7"} Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.591481 4960 scope.go:117] "RemoveContainer" containerID="b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.591656 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-757kx" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.627292 4960 scope.go:117] "RemoveContainer" containerID="142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.651141 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-757kx"] Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.653149 4960 scope.go:117] "RemoveContainer" containerID="6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.674574 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-757kx"] Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.710174 4960 scope.go:117] "RemoveContainer" containerID="b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b" Oct 02 08:25:10 crc kubenswrapper[4960]: E1002 08:25:10.711150 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b\": container with ID starting with b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b not found: ID does not exist" containerID="b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.711211 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b"} err="failed to get container status \"b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b\": rpc error: code = NotFound desc = could not find container \"b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b\": container with ID starting with b8ef041b6626dc97667d6ae61bff9dee691070a34b0ee2e81b1f55f6a5fa448b not found: ID does not exist" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.711256 4960 scope.go:117] "RemoveContainer" containerID="142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098" Oct 02 08:25:10 crc kubenswrapper[4960]: E1002 08:25:10.711727 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098\": container with ID starting with 142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098 not found: ID does not exist" containerID="142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.711862 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098"} err="failed to get container status \"142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098\": rpc error: code = NotFound desc = could not find container \"142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098\": container with ID starting with 142d1a2e7ec77de46e1a33eea2d24c65ad16830dd2edd197b0f792648802d098 not found: ID does not exist" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.712016 4960 scope.go:117] "RemoveContainer" containerID="6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b" Oct 02 08:25:10 crc kubenswrapper[4960]: E1002 08:25:10.712611 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b\": container with ID starting with 6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b not found: ID does not exist" containerID="6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b" Oct 02 08:25:10 crc kubenswrapper[4960]: I1002 08:25:10.712741 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b"} err="failed to get container status \"6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b\": rpc error: code = NotFound desc = could not find container \"6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b\": container with ID starting with 6d3fba468077358df7f2e5c5110b1567cbf88ac9ab01c7c4ca6d24212afec95b not found: ID does not exist" Oct 02 08:25:12 crc kubenswrapper[4960]: I1002 08:25:12.348510 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" path="/var/lib/kubelet/pods/c0b733a5-a8da-4e7e-a08e-18412f72cc1b/volumes" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.459618 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-chpn9"] Oct 02 08:26:31 crc kubenswrapper[4960]: E1002 08:26:31.461014 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="extract-content" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.461033 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="extract-content" Oct 02 08:26:31 crc kubenswrapper[4960]: E1002 08:26:31.461054 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="extract-utilities" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.461061 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="extract-utilities" Oct 02 08:26:31 crc kubenswrapper[4960]: E1002 08:26:31.461082 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="registry-server" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.461089 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="registry-server" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.461313 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b733a5-a8da-4e7e-a08e-18412f72cc1b" containerName="registry-server" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.463013 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.475866 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-chpn9"] Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.537414 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-catalog-content\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.537493 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srfl7\" (UniqueName: \"kubernetes.io/projected/5ea67092-6947-40e8-80ee-28978fd01109-kube-api-access-srfl7\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.537673 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-utilities\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.639964 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-catalog-content\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.640085 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srfl7\" (UniqueName: \"kubernetes.io/projected/5ea67092-6947-40e8-80ee-28978fd01109-kube-api-access-srfl7\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.640250 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-utilities\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.640602 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-catalog-content\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.640762 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-utilities\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.658519 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srfl7\" (UniqueName: \"kubernetes.io/projected/5ea67092-6947-40e8-80ee-28978fd01109-kube-api-access-srfl7\") pod \"community-operators-chpn9\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:31 crc kubenswrapper[4960]: I1002 08:26:31.789770 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:32 crc kubenswrapper[4960]: I1002 08:26:32.379822 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-chpn9"] Oct 02 08:26:33 crc kubenswrapper[4960]: I1002 08:26:33.391911 4960 generic.go:334] "Generic (PLEG): container finished" podID="5ea67092-6947-40e8-80ee-28978fd01109" containerID="46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c" exitCode=0 Oct 02 08:26:33 crc kubenswrapper[4960]: I1002 08:26:33.392019 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chpn9" event={"ID":"5ea67092-6947-40e8-80ee-28978fd01109","Type":"ContainerDied","Data":"46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c"} Oct 02 08:26:33 crc kubenswrapper[4960]: I1002 08:26:33.392388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chpn9" event={"ID":"5ea67092-6947-40e8-80ee-28978fd01109","Type":"ContainerStarted","Data":"19a63a0b0556670609d4a34f500adbec7913a23a8d15449deacc74f18d6e50af"} Oct 02 08:26:33 crc kubenswrapper[4960]: I1002 08:26:33.397168 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:26:35 crc kubenswrapper[4960]: I1002 08:26:35.416534 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chpn9" event={"ID":"5ea67092-6947-40e8-80ee-28978fd01109","Type":"ContainerStarted","Data":"accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66"} Oct 02 08:26:36 crc kubenswrapper[4960]: I1002 08:26:36.430828 4960 generic.go:334] "Generic (PLEG): container finished" podID="5ea67092-6947-40e8-80ee-28978fd01109" containerID="accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66" exitCode=0 Oct 02 08:26:36 crc kubenswrapper[4960]: I1002 08:26:36.431111 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chpn9" event={"ID":"5ea67092-6947-40e8-80ee-28978fd01109","Type":"ContainerDied","Data":"accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66"} Oct 02 08:26:37 crc kubenswrapper[4960]: I1002 08:26:37.453641 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chpn9" event={"ID":"5ea67092-6947-40e8-80ee-28978fd01109","Type":"ContainerStarted","Data":"edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f"} Oct 02 08:26:37 crc kubenswrapper[4960]: I1002 08:26:37.499094 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-chpn9" podStartSLOduration=3.024409762 podStartE2EDuration="6.499069602s" podCreationTimestamp="2025-10-02 08:26:31 +0000 UTC" firstStartedPulling="2025-10-02 08:26:33.39696365 +0000 UTC m=+4214.428909937" lastFinishedPulling="2025-10-02 08:26:36.87162349 +0000 UTC m=+4217.903569777" observedRunningTime="2025-10-02 08:26:37.476806176 +0000 UTC m=+4218.508752463" watchObservedRunningTime="2025-10-02 08:26:37.499069602 +0000 UTC m=+4218.531015889" Oct 02 08:26:41 crc kubenswrapper[4960]: I1002 08:26:41.790178 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:41 crc kubenswrapper[4960]: I1002 08:26:41.791004 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:42 crc kubenswrapper[4960]: I1002 08:26:42.447582 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:42 crc kubenswrapper[4960]: I1002 08:26:42.565455 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:42 crc kubenswrapper[4960]: I1002 08:26:42.691141 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-chpn9"] Oct 02 08:26:44 crc kubenswrapper[4960]: I1002 08:26:44.526251 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-chpn9" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="registry-server" containerID="cri-o://edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f" gracePeriod=2 Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.231271 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.295647 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srfl7\" (UniqueName: \"kubernetes.io/projected/5ea67092-6947-40e8-80ee-28978fd01109-kube-api-access-srfl7\") pod \"5ea67092-6947-40e8-80ee-28978fd01109\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.296258 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-catalog-content\") pod \"5ea67092-6947-40e8-80ee-28978fd01109\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.296635 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-utilities\") pod \"5ea67092-6947-40e8-80ee-28978fd01109\" (UID: \"5ea67092-6947-40e8-80ee-28978fd01109\") " Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.298051 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-utilities" (OuterVolumeSpecName: "utilities") pod "5ea67092-6947-40e8-80ee-28978fd01109" (UID: "5ea67092-6947-40e8-80ee-28978fd01109"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.298920 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.323839 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ea67092-6947-40e8-80ee-28978fd01109-kube-api-access-srfl7" (OuterVolumeSpecName: "kube-api-access-srfl7") pod "5ea67092-6947-40e8-80ee-28978fd01109" (UID: "5ea67092-6947-40e8-80ee-28978fd01109"). InnerVolumeSpecName "kube-api-access-srfl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.402815 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srfl7\" (UniqueName: \"kubernetes.io/projected/5ea67092-6947-40e8-80ee-28978fd01109-kube-api-access-srfl7\") on node \"crc\" DevicePath \"\"" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.541168 4960 generic.go:334] "Generic (PLEG): container finished" podID="5ea67092-6947-40e8-80ee-28978fd01109" containerID="edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f" exitCode=0 Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.541214 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chpn9" event={"ID":"5ea67092-6947-40e8-80ee-28978fd01109","Type":"ContainerDied","Data":"edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f"} Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.541242 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chpn9" event={"ID":"5ea67092-6947-40e8-80ee-28978fd01109","Type":"ContainerDied","Data":"19a63a0b0556670609d4a34f500adbec7913a23a8d15449deacc74f18d6e50af"} Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.541264 4960 scope.go:117] "RemoveContainer" containerID="edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.541264 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chpn9" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.563278 4960 scope.go:117] "RemoveContainer" containerID="accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.590585 4960 scope.go:117] "RemoveContainer" containerID="46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.632918 4960 scope.go:117] "RemoveContainer" containerID="edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f" Oct 02 08:26:45 crc kubenswrapper[4960]: E1002 08:26:45.633564 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f\": container with ID starting with edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f not found: ID does not exist" containerID="edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.633599 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f"} err="failed to get container status \"edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f\": rpc error: code = NotFound desc = could not find container \"edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f\": container with ID starting with edf26609ec0fce7770c54ae71004e3db63c6df1cf0bf9c370de5e1689cab298f not found: ID does not exist" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.633621 4960 scope.go:117] "RemoveContainer" containerID="accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66" Oct 02 08:26:45 crc kubenswrapper[4960]: E1002 08:26:45.633854 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66\": container with ID starting with accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66 not found: ID does not exist" containerID="accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.633871 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66"} err="failed to get container status \"accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66\": rpc error: code = NotFound desc = could not find container \"accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66\": container with ID starting with accb22248fb2862fc0385505a7ae2981ec205d6ec3f95827544fe05b31962b66 not found: ID does not exist" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.633883 4960 scope.go:117] "RemoveContainer" containerID="46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c" Oct 02 08:26:45 crc kubenswrapper[4960]: E1002 08:26:45.634123 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c\": container with ID starting with 46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c not found: ID does not exist" containerID="46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.634140 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c"} err="failed to get container status \"46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c\": rpc error: code = NotFound desc = could not find container \"46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c\": container with ID starting with 46f43d512ea6e374cd9b42661fdb49fd1ced25b6be2e8026f51f5e3bb781166c not found: ID does not exist" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.854824 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ea67092-6947-40e8-80ee-28978fd01109" (UID: "5ea67092-6947-40e8-80ee-28978fd01109"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:26:45 crc kubenswrapper[4960]: I1002 08:26:45.915339 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea67092-6947-40e8-80ee-28978fd01109-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:26:46 crc kubenswrapper[4960]: I1002 08:26:46.211358 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-chpn9"] Oct 02 08:26:46 crc kubenswrapper[4960]: I1002 08:26:46.233467 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-chpn9"] Oct 02 08:26:46 crc kubenswrapper[4960]: I1002 08:26:46.344584 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ea67092-6947-40e8-80ee-28978fd01109" path="/var/lib/kubelet/pods/5ea67092-6947-40e8-80ee-28978fd01109/volumes" Oct 02 08:26:59 crc kubenswrapper[4960]: I1002 08:26:59.150414 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:26:59 crc kubenswrapper[4960]: I1002 08:26:59.152389 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:27:29 crc kubenswrapper[4960]: I1002 08:27:29.150517 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:27:29 crc kubenswrapper[4960]: I1002 08:27:29.151383 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.149739 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.150489 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.150546 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.151417 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.151482 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" gracePeriod=600 Oct 02 08:27:59 crc kubenswrapper[4960]: E1002 08:27:59.290672 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.343243 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" exitCode=0 Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.343327 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056"} Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.343385 4960 scope.go:117] "RemoveContainer" containerID="e78d5afcfa8229fda3c1a41a2550cb021603e250d345d7dbd2122e8b207bc886" Oct 02 08:27:59 crc kubenswrapper[4960]: I1002 08:27:59.344568 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:27:59 crc kubenswrapper[4960]: E1002 08:27:59.345129 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:28:11 crc kubenswrapper[4960]: I1002 08:28:11.332283 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:28:11 crc kubenswrapper[4960]: E1002 08:28:11.333027 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:28:26 crc kubenswrapper[4960]: I1002 08:28:26.330725 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:28:26 crc kubenswrapper[4960]: E1002 08:28:26.332071 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:28:40 crc kubenswrapper[4960]: I1002 08:28:40.337527 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:28:40 crc kubenswrapper[4960]: E1002 08:28:40.338961 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:28:51 crc kubenswrapper[4960]: I1002 08:28:51.330330 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:28:51 crc kubenswrapper[4960]: E1002 08:28:51.331128 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:29:02 crc kubenswrapper[4960]: I1002 08:29:02.330824 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:29:02 crc kubenswrapper[4960]: E1002 08:29:02.331934 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:29:17 crc kubenswrapper[4960]: I1002 08:29:17.330592 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:29:17 crc kubenswrapper[4960]: E1002 08:29:17.332337 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:29:32 crc kubenswrapper[4960]: I1002 08:29:32.330479 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:29:32 crc kubenswrapper[4960]: E1002 08:29:32.331454 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:29:43 crc kubenswrapper[4960]: I1002 08:29:43.330451 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:29:43 crc kubenswrapper[4960]: E1002 08:29:43.331291 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:29:57 crc kubenswrapper[4960]: I1002 08:29:57.331366 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:29:57 crc kubenswrapper[4960]: E1002 08:29:57.332629 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.141958 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv"] Oct 02 08:30:00 crc kubenswrapper[4960]: E1002 08:30:00.143327 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="extract-utilities" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.143346 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="extract-utilities" Oct 02 08:30:00 crc kubenswrapper[4960]: E1002 08:30:00.143369 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="registry-server" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.143377 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="registry-server" Oct 02 08:30:00 crc kubenswrapper[4960]: E1002 08:30:00.143418 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="extract-content" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.143428 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="extract-content" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.143666 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea67092-6947-40e8-80ee-28978fd01109" containerName="registry-server" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.144669 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.146530 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.147418 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.159680 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv"] Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.318103 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89s2g\" (UniqueName: \"kubernetes.io/projected/d11f2c48-b7f7-4106-aedc-e2b94cab2798-kube-api-access-89s2g\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.318165 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d11f2c48-b7f7-4106-aedc-e2b94cab2798-secret-volume\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.318236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d11f2c48-b7f7-4106-aedc-e2b94cab2798-config-volume\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.421180 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89s2g\" (UniqueName: \"kubernetes.io/projected/d11f2c48-b7f7-4106-aedc-e2b94cab2798-kube-api-access-89s2g\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.421272 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d11f2c48-b7f7-4106-aedc-e2b94cab2798-secret-volume\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.421377 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d11f2c48-b7f7-4106-aedc-e2b94cab2798-config-volume\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.422512 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d11f2c48-b7f7-4106-aedc-e2b94cab2798-config-volume\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.428212 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d11f2c48-b7f7-4106-aedc-e2b94cab2798-secret-volume\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.440916 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89s2g\" (UniqueName: \"kubernetes.io/projected/d11f2c48-b7f7-4106-aedc-e2b94cab2798-kube-api-access-89s2g\") pod \"collect-profiles-29323230-n45xv\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:00 crc kubenswrapper[4960]: I1002 08:30:00.502441 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:01 crc kubenswrapper[4960]: I1002 08:30:01.017826 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv"] Oct 02 08:30:01 crc kubenswrapper[4960]: I1002 08:30:01.767225 4960 generic.go:334] "Generic (PLEG): container finished" podID="d11f2c48-b7f7-4106-aedc-e2b94cab2798" containerID="0d7ac2e1c12432571509cdd04e63c9c47b4e2c4ef0a546fae51861ecbec23b68" exitCode=0 Oct 02 08:30:01 crc kubenswrapper[4960]: I1002 08:30:01.767544 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" event={"ID":"d11f2c48-b7f7-4106-aedc-e2b94cab2798","Type":"ContainerDied","Data":"0d7ac2e1c12432571509cdd04e63c9c47b4e2c4ef0a546fae51861ecbec23b68"} Oct 02 08:30:01 crc kubenswrapper[4960]: I1002 08:30:01.769320 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" event={"ID":"d11f2c48-b7f7-4106-aedc-e2b94cab2798","Type":"ContainerStarted","Data":"db2edb46d3b188e434302c47b788dbbe18fc89df26feb2ce11468ec8d40e0adf"} Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.465253 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.593524 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89s2g\" (UniqueName: \"kubernetes.io/projected/d11f2c48-b7f7-4106-aedc-e2b94cab2798-kube-api-access-89s2g\") pod \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.593609 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d11f2c48-b7f7-4106-aedc-e2b94cab2798-secret-volume\") pod \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.593715 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d11f2c48-b7f7-4106-aedc-e2b94cab2798-config-volume\") pod \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\" (UID: \"d11f2c48-b7f7-4106-aedc-e2b94cab2798\") " Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.594476 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d11f2c48-b7f7-4106-aedc-e2b94cab2798-config-volume" (OuterVolumeSpecName: "config-volume") pod "d11f2c48-b7f7-4106-aedc-e2b94cab2798" (UID: "d11f2c48-b7f7-4106-aedc-e2b94cab2798"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.601347 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d11f2c48-b7f7-4106-aedc-e2b94cab2798-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d11f2c48-b7f7-4106-aedc-e2b94cab2798" (UID: "d11f2c48-b7f7-4106-aedc-e2b94cab2798"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.609227 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d11f2c48-b7f7-4106-aedc-e2b94cab2798-kube-api-access-89s2g" (OuterVolumeSpecName: "kube-api-access-89s2g") pod "d11f2c48-b7f7-4106-aedc-e2b94cab2798" (UID: "d11f2c48-b7f7-4106-aedc-e2b94cab2798"). InnerVolumeSpecName "kube-api-access-89s2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.696813 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89s2g\" (UniqueName: \"kubernetes.io/projected/d11f2c48-b7f7-4106-aedc-e2b94cab2798-kube-api-access-89s2g\") on node \"crc\" DevicePath \"\"" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.696851 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d11f2c48-b7f7-4106-aedc-e2b94cab2798-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.696861 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d11f2c48-b7f7-4106-aedc-e2b94cab2798-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.790308 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" event={"ID":"d11f2c48-b7f7-4106-aedc-e2b94cab2798","Type":"ContainerDied","Data":"db2edb46d3b188e434302c47b788dbbe18fc89df26feb2ce11468ec8d40e0adf"} Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.790351 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db2edb46d3b188e434302c47b788dbbe18fc89df26feb2ce11468ec8d40e0adf" Oct 02 08:30:03 crc kubenswrapper[4960]: I1002 08:30:03.790401 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv" Oct 02 08:30:04 crc kubenswrapper[4960]: I1002 08:30:04.537787 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt"] Oct 02 08:30:04 crc kubenswrapper[4960]: I1002 08:30:04.546233 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-lfrzt"] Oct 02 08:30:06 crc kubenswrapper[4960]: I1002 08:30:06.344906 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f" path="/var/lib/kubelet/pods/c2dddc57-7b2c-431a-b3b4-c0f0680a1b9f/volumes" Oct 02 08:30:09 crc kubenswrapper[4960]: I1002 08:30:09.330172 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:30:09 crc kubenswrapper[4960]: E1002 08:30:09.330964 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:30:22 crc kubenswrapper[4960]: I1002 08:30:22.335061 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:30:22 crc kubenswrapper[4960]: E1002 08:30:22.336452 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:30:37 crc kubenswrapper[4960]: I1002 08:30:37.330725 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:30:37 crc kubenswrapper[4960]: E1002 08:30:37.332050 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:30:49 crc kubenswrapper[4960]: I1002 08:30:49.331052 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:30:49 crc kubenswrapper[4960]: E1002 08:30:49.332059 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:31:02 crc kubenswrapper[4960]: I1002 08:31:02.882308 4960 scope.go:117] "RemoveContainer" containerID="fb2feb7439a262cbba4173100dc1ae7ccb25788d5e18bc001697bb1457c4fa70" Oct 02 08:31:03 crc kubenswrapper[4960]: I1002 08:31:03.330470 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:31:03 crc kubenswrapper[4960]: E1002 08:31:03.331935 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:31:15 crc kubenswrapper[4960]: I1002 08:31:15.330882 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:31:15 crc kubenswrapper[4960]: E1002 08:31:15.332555 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:31:27 crc kubenswrapper[4960]: I1002 08:31:27.331396 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:31:27 crc kubenswrapper[4960]: E1002 08:31:27.332392 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:31:41 crc kubenswrapper[4960]: I1002 08:31:41.330951 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:31:41 crc kubenswrapper[4960]: E1002 08:31:41.332709 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:31:55 crc kubenswrapper[4960]: I1002 08:31:55.330673 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:31:55 crc kubenswrapper[4960]: E1002 08:31:55.332776 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:32:09 crc kubenswrapper[4960]: I1002 08:32:09.330454 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:32:09 crc kubenswrapper[4960]: E1002 08:32:09.331795 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:32:22 crc kubenswrapper[4960]: I1002 08:32:22.330886 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:32:22 crc kubenswrapper[4960]: E1002 08:32:22.331924 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.048773 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vbm6z"] Oct 02 08:32:23 crc kubenswrapper[4960]: E1002 08:32:23.050704 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d11f2c48-b7f7-4106-aedc-e2b94cab2798" containerName="collect-profiles" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.050726 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d11f2c48-b7f7-4106-aedc-e2b94cab2798" containerName="collect-profiles" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.051096 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d11f2c48-b7f7-4106-aedc-e2b94cab2798" containerName="collect-profiles" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.057345 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.068016 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbm6z"] Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.253169 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbwg9\" (UniqueName: \"kubernetes.io/projected/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-kube-api-access-lbwg9\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.253532 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-utilities\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.253585 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-catalog-content\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.356285 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbwg9\" (UniqueName: \"kubernetes.io/projected/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-kube-api-access-lbwg9\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.356429 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-utilities\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.356486 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-catalog-content\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.357046 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-utilities\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.357134 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-catalog-content\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.399226 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbwg9\" (UniqueName: \"kubernetes.io/projected/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-kube-api-access-lbwg9\") pod \"redhat-marketplace-vbm6z\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:23 crc kubenswrapper[4960]: I1002 08:32:23.684688 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:24 crc kubenswrapper[4960]: I1002 08:32:24.201460 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbm6z"] Oct 02 08:32:24 crc kubenswrapper[4960]: I1002 08:32:24.267428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbm6z" event={"ID":"f65d1463-3f4d-49a1-9933-a06a3c2b4a35","Type":"ContainerStarted","Data":"2e236d1ace5322a93d3e888f941e8788a3cee6f8e96aba0d8077e45f29dea035"} Oct 02 08:32:25 crc kubenswrapper[4960]: I1002 08:32:25.281036 4960 generic.go:334] "Generic (PLEG): container finished" podID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerID="2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6" exitCode=0 Oct 02 08:32:25 crc kubenswrapper[4960]: I1002 08:32:25.281482 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbm6z" event={"ID":"f65d1463-3f4d-49a1-9933-a06a3c2b4a35","Type":"ContainerDied","Data":"2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6"} Oct 02 08:32:25 crc kubenswrapper[4960]: I1002 08:32:25.284712 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:32:27 crc kubenswrapper[4960]: I1002 08:32:27.320865 4960 generic.go:334] "Generic (PLEG): container finished" podID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerID="59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4" exitCode=0 Oct 02 08:32:27 crc kubenswrapper[4960]: I1002 08:32:27.321024 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbm6z" event={"ID":"f65d1463-3f4d-49a1-9933-a06a3c2b4a35","Type":"ContainerDied","Data":"59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4"} Oct 02 08:32:28 crc kubenswrapper[4960]: I1002 08:32:28.341827 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbm6z" event={"ID":"f65d1463-3f4d-49a1-9933-a06a3c2b4a35","Type":"ContainerStarted","Data":"55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74"} Oct 02 08:32:28 crc kubenswrapper[4960]: I1002 08:32:28.368086 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vbm6z" podStartSLOduration=2.835530951 podStartE2EDuration="5.36796924s" podCreationTimestamp="2025-10-02 08:32:23 +0000 UTC" firstStartedPulling="2025-10-02 08:32:25.2842566 +0000 UTC m=+4566.316202907" lastFinishedPulling="2025-10-02 08:32:27.816694909 +0000 UTC m=+4568.848641196" observedRunningTime="2025-10-02 08:32:28.363318244 +0000 UTC m=+4569.395264541" watchObservedRunningTime="2025-10-02 08:32:28.36796924 +0000 UTC m=+4569.399915537" Oct 02 08:32:33 crc kubenswrapper[4960]: I1002 08:32:33.685763 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:33 crc kubenswrapper[4960]: I1002 08:32:33.687432 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:33 crc kubenswrapper[4960]: I1002 08:32:33.746670 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:34 crc kubenswrapper[4960]: I1002 08:32:34.455868 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:34 crc kubenswrapper[4960]: I1002 08:32:34.514352 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbm6z"] Oct 02 08:32:35 crc kubenswrapper[4960]: I1002 08:32:35.330224 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:32:35 crc kubenswrapper[4960]: E1002 08:32:35.330567 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:32:36 crc kubenswrapper[4960]: I1002 08:32:36.424918 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vbm6z" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="registry-server" containerID="cri-o://55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74" gracePeriod=2 Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.136172 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.315239 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-catalog-content\") pod \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.315530 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-utilities\") pod \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.315658 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbwg9\" (UniqueName: \"kubernetes.io/projected/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-kube-api-access-lbwg9\") pod \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\" (UID: \"f65d1463-3f4d-49a1-9933-a06a3c2b4a35\") " Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.316845 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-utilities" (OuterVolumeSpecName: "utilities") pod "f65d1463-3f4d-49a1-9933-a06a3c2b4a35" (UID: "f65d1463-3f4d-49a1-9933-a06a3c2b4a35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.325503 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-kube-api-access-lbwg9" (OuterVolumeSpecName: "kube-api-access-lbwg9") pod "f65d1463-3f4d-49a1-9933-a06a3c2b4a35" (UID: "f65d1463-3f4d-49a1-9933-a06a3c2b4a35"). InnerVolumeSpecName "kube-api-access-lbwg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.330898 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f65d1463-3f4d-49a1-9933-a06a3c2b4a35" (UID: "f65d1463-3f4d-49a1-9933-a06a3c2b4a35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.422509 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.423107 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbwg9\" (UniqueName: \"kubernetes.io/projected/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-kube-api-access-lbwg9\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.423126 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65d1463-3f4d-49a1-9933-a06a3c2b4a35-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.438362 4960 generic.go:334] "Generic (PLEG): container finished" podID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerID="55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74" exitCode=0 Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.438420 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbm6z" event={"ID":"f65d1463-3f4d-49a1-9933-a06a3c2b4a35","Type":"ContainerDied","Data":"55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74"} Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.438453 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbm6z" event={"ID":"f65d1463-3f4d-49a1-9933-a06a3c2b4a35","Type":"ContainerDied","Data":"2e236d1ace5322a93d3e888f941e8788a3cee6f8e96aba0d8077e45f29dea035"} Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.438471 4960 scope.go:117] "RemoveContainer" containerID="55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.438641 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbm6z" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.487019 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbm6z"] Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.494675 4960 scope.go:117] "RemoveContainer" containerID="59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.498550 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbm6z"] Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.515907 4960 scope.go:117] "RemoveContainer" containerID="2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.568670 4960 scope.go:117] "RemoveContainer" containerID="55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74" Oct 02 08:32:37 crc kubenswrapper[4960]: E1002 08:32:37.569485 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74\": container with ID starting with 55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74 not found: ID does not exist" containerID="55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.569545 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74"} err="failed to get container status \"55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74\": rpc error: code = NotFound desc = could not find container \"55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74\": container with ID starting with 55af2bc971ad91920d4e25cbca2f027ad5ba92c885a1ac4734908a4971b25f74 not found: ID does not exist" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.569611 4960 scope.go:117] "RemoveContainer" containerID="59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4" Oct 02 08:32:37 crc kubenswrapper[4960]: E1002 08:32:37.570179 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4\": container with ID starting with 59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4 not found: ID does not exist" containerID="59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.570268 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4"} err="failed to get container status \"59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4\": rpc error: code = NotFound desc = could not find container \"59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4\": container with ID starting with 59996b4d54ca19ffa79cb83f43c3330fe8904842cab5256ae7f8330eb68b82b4 not found: ID does not exist" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.570312 4960 scope.go:117] "RemoveContainer" containerID="2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6" Oct 02 08:32:37 crc kubenswrapper[4960]: E1002 08:32:37.570812 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6\": container with ID starting with 2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6 not found: ID does not exist" containerID="2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6" Oct 02 08:32:37 crc kubenswrapper[4960]: I1002 08:32:37.570868 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6"} err="failed to get container status \"2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6\": rpc error: code = NotFound desc = could not find container \"2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6\": container with ID starting with 2689da68182cdbddc6180ec74def9b92b9efeee82b0af0071633898921b455d6 not found: ID does not exist" Oct 02 08:32:38 crc kubenswrapper[4960]: I1002 08:32:38.345309 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" path="/var/lib/kubelet/pods/f65d1463-3f4d-49a1-9933-a06a3c2b4a35/volumes" Oct 02 08:32:46 crc kubenswrapper[4960]: I1002 08:32:46.330323 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:32:46 crc kubenswrapper[4960]: E1002 08:32:46.331141 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:32:58 crc kubenswrapper[4960]: I1002 08:32:58.330736 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:32:58 crc kubenswrapper[4960]: E1002 08:32:58.331690 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:33:09 crc kubenswrapper[4960]: I1002 08:33:09.330617 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:33:09 crc kubenswrapper[4960]: I1002 08:33:09.776413 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"82652257eb990b346f43463872669f28c29971e7727a85fb74114ad3771937fe"} Oct 02 08:35:29 crc kubenswrapper[4960]: I1002 08:35:29.150394 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:35:29 crc kubenswrapper[4960]: I1002 08:35:29.151130 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.072720 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dbk57"] Oct 02 08:35:52 crc kubenswrapper[4960]: E1002 08:35:52.074068 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="extract-content" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.074090 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="extract-content" Oct 02 08:35:52 crc kubenswrapper[4960]: E1002 08:35:52.074108 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="registry-server" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.074117 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="registry-server" Oct 02 08:35:52 crc kubenswrapper[4960]: E1002 08:35:52.074145 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="extract-utilities" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.074154 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="extract-utilities" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.074427 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f65d1463-3f4d-49a1-9933-a06a3c2b4a35" containerName="registry-server" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.076426 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.086631 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dbk57"] Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.205525 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-utilities\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.205724 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-catalog-content\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.205788 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d8zs\" (UniqueName: \"kubernetes.io/projected/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-kube-api-access-6d8zs\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.308072 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-catalog-content\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.308152 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d8zs\" (UniqueName: \"kubernetes.io/projected/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-kube-api-access-6d8zs\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.308235 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-utilities\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.308712 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-catalog-content\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.308743 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-utilities\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.337822 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d8zs\" (UniqueName: \"kubernetes.io/projected/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-kube-api-access-6d8zs\") pod \"redhat-operators-dbk57\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.431742 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:35:52 crc kubenswrapper[4960]: I1002 08:35:52.935339 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dbk57"] Oct 02 08:35:53 crc kubenswrapper[4960]: I1002 08:35:53.501183 4960 generic.go:334] "Generic (PLEG): container finished" podID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerID="3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938" exitCode=0 Oct 02 08:35:53 crc kubenswrapper[4960]: I1002 08:35:53.501299 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbk57" event={"ID":"9a0834aa-904c-4e7a-a176-0ce2d5d2321f","Type":"ContainerDied","Data":"3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938"} Oct 02 08:35:53 crc kubenswrapper[4960]: I1002 08:35:53.501660 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbk57" event={"ID":"9a0834aa-904c-4e7a-a176-0ce2d5d2321f","Type":"ContainerStarted","Data":"a9aefc562ff4cde5b8c07fdd995e249692d9dbf4fd01f51b652abe09ea7bfc45"} Oct 02 08:35:55 crc kubenswrapper[4960]: I1002 08:35:55.524401 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbk57" event={"ID":"9a0834aa-904c-4e7a-a176-0ce2d5d2321f","Type":"ContainerStarted","Data":"af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876"} Oct 02 08:35:56 crc kubenswrapper[4960]: I1002 08:35:56.542203 4960 generic.go:334] "Generic (PLEG): container finished" podID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerID="af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876" exitCode=0 Oct 02 08:35:56 crc kubenswrapper[4960]: I1002 08:35:56.542313 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbk57" event={"ID":"9a0834aa-904c-4e7a-a176-0ce2d5d2321f","Type":"ContainerDied","Data":"af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876"} Oct 02 08:35:57 crc kubenswrapper[4960]: I1002 08:35:57.553099 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbk57" event={"ID":"9a0834aa-904c-4e7a-a176-0ce2d5d2321f","Type":"ContainerStarted","Data":"5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3"} Oct 02 08:35:57 crc kubenswrapper[4960]: I1002 08:35:57.570489 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dbk57" podStartSLOduration=1.935048917 podStartE2EDuration="5.570467988s" podCreationTimestamp="2025-10-02 08:35:52 +0000 UTC" firstStartedPulling="2025-10-02 08:35:53.504261432 +0000 UTC m=+4774.536207719" lastFinishedPulling="2025-10-02 08:35:57.139680503 +0000 UTC m=+4778.171626790" observedRunningTime="2025-10-02 08:35:57.567771006 +0000 UTC m=+4778.599717293" watchObservedRunningTime="2025-10-02 08:35:57.570467988 +0000 UTC m=+4778.602414275" Oct 02 08:35:59 crc kubenswrapper[4960]: I1002 08:35:59.149733 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:35:59 crc kubenswrapper[4960]: I1002 08:35:59.151269 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:36:02 crc kubenswrapper[4960]: I1002 08:36:02.432664 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:36:02 crc kubenswrapper[4960]: I1002 08:36:02.433496 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:36:03 crc kubenswrapper[4960]: I1002 08:36:03.506235 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dbk57" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="registry-server" probeResult="failure" output=< Oct 02 08:36:03 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 08:36:03 crc kubenswrapper[4960]: > Oct 02 08:36:12 crc kubenswrapper[4960]: I1002 08:36:12.563942 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:36:12 crc kubenswrapper[4960]: I1002 08:36:12.635502 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:36:12 crc kubenswrapper[4960]: I1002 08:36:12.811055 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dbk57"] Oct 02 08:36:13 crc kubenswrapper[4960]: I1002 08:36:13.701859 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dbk57" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="registry-server" containerID="cri-o://5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3" gracePeriod=2 Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.303245 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.433121 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-catalog-content\") pod \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.433293 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-utilities\") pod \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.433367 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d8zs\" (UniqueName: \"kubernetes.io/projected/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-kube-api-access-6d8zs\") pod \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\" (UID: \"9a0834aa-904c-4e7a-a176-0ce2d5d2321f\") " Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.434564 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-utilities" (OuterVolumeSpecName: "utilities") pod "9a0834aa-904c-4e7a-a176-0ce2d5d2321f" (UID: "9a0834aa-904c-4e7a-a176-0ce2d5d2321f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.446289 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-kube-api-access-6d8zs" (OuterVolumeSpecName: "kube-api-access-6d8zs") pod "9a0834aa-904c-4e7a-a176-0ce2d5d2321f" (UID: "9a0834aa-904c-4e7a-a176-0ce2d5d2321f"). InnerVolumeSpecName "kube-api-access-6d8zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.518615 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a0834aa-904c-4e7a-a176-0ce2d5d2321f" (UID: "9a0834aa-904c-4e7a-a176-0ce2d5d2321f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.536345 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.536379 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.536391 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d8zs\" (UniqueName: \"kubernetes.io/projected/9a0834aa-904c-4e7a-a176-0ce2d5d2321f-kube-api-access-6d8zs\") on node \"crc\" DevicePath \"\"" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.714306 4960 generic.go:334] "Generic (PLEG): container finished" podID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerID="5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3" exitCode=0 Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.714542 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbk57" event={"ID":"9a0834aa-904c-4e7a-a176-0ce2d5d2321f","Type":"ContainerDied","Data":"5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3"} Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.714736 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbk57" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.714766 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbk57" event={"ID":"9a0834aa-904c-4e7a-a176-0ce2d5d2321f","Type":"ContainerDied","Data":"a9aefc562ff4cde5b8c07fdd995e249692d9dbf4fd01f51b652abe09ea7bfc45"} Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.714795 4960 scope.go:117] "RemoveContainer" containerID="5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.739291 4960 scope.go:117] "RemoveContainer" containerID="af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.769276 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dbk57"] Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.778225 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dbk57"] Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.798408 4960 scope.go:117] "RemoveContainer" containerID="3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.829829 4960 scope.go:117] "RemoveContainer" containerID="5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3" Oct 02 08:36:14 crc kubenswrapper[4960]: E1002 08:36:14.830502 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3\": container with ID starting with 5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3 not found: ID does not exist" containerID="5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.830547 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3"} err="failed to get container status \"5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3\": rpc error: code = NotFound desc = could not find container \"5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3\": container with ID starting with 5413c5be82e0939eb3e4e6caf3c268176a8febbc99de90744169846a8f4ee0b3 not found: ID does not exist" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.830576 4960 scope.go:117] "RemoveContainer" containerID="af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876" Oct 02 08:36:14 crc kubenswrapper[4960]: E1002 08:36:14.831043 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876\": container with ID starting with af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876 not found: ID does not exist" containerID="af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.831075 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876"} err="failed to get container status \"af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876\": rpc error: code = NotFound desc = could not find container \"af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876\": container with ID starting with af0ab02e84c8961e3b263fcdb49f12e5c5f546aabb2b3cc598a5fd4615c1c876 not found: ID does not exist" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.831093 4960 scope.go:117] "RemoveContainer" containerID="3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938" Oct 02 08:36:14 crc kubenswrapper[4960]: E1002 08:36:14.831461 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938\": container with ID starting with 3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938 not found: ID does not exist" containerID="3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938" Oct 02 08:36:14 crc kubenswrapper[4960]: I1002 08:36:14.831491 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938"} err="failed to get container status \"3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938\": rpc error: code = NotFound desc = could not find container \"3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938\": container with ID starting with 3baa2aec7dfac4abc4a33387fb8e5bf8c82c3e1e68143242de8e4eadc1c2a938 not found: ID does not exist" Oct 02 08:36:16 crc kubenswrapper[4960]: I1002 08:36:16.345158 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" path="/var/lib/kubelet/pods/9a0834aa-904c-4e7a-a176-0ce2d5d2321f/volumes" Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.150734 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.152097 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.152196 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.153681 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"82652257eb990b346f43463872669f28c29971e7727a85fb74114ad3771937fe"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.153804 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://82652257eb990b346f43463872669f28c29971e7727a85fb74114ad3771937fe" gracePeriod=600 Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.935551 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="82652257eb990b346f43463872669f28c29971e7727a85fb74114ad3771937fe" exitCode=0 Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.935620 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"82652257eb990b346f43463872669f28c29971e7727a85fb74114ad3771937fe"} Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.936176 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f"} Oct 02 08:36:29 crc kubenswrapper[4960]: I1002 08:36:29.936201 4960 scope.go:117] "RemoveContainer" containerID="bcc8bfce0f74546e6c57d713ee9e5b0f852d73130fc5ddcddda03ce471021056" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.150909 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cjskg"] Oct 02 08:36:59 crc kubenswrapper[4960]: E1002 08:36:59.152159 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="extract-content" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.152175 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="extract-content" Oct 02 08:36:59 crc kubenswrapper[4960]: E1002 08:36:59.152189 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="extract-utilities" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.152197 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="extract-utilities" Oct 02 08:36:59 crc kubenswrapper[4960]: E1002 08:36:59.152226 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="registry-server" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.152233 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="registry-server" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.152422 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0834aa-904c-4e7a-a176-0ce2d5d2321f" containerName="registry-server" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.153804 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.190437 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjskg"] Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.268455 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrqs9\" (UniqueName: \"kubernetes.io/projected/b72e6503-3f54-43d9-ae0f-0240e1411010-kube-api-access-qrqs9\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.268568 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-utilities\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.268672 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-catalog-content\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.370795 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrqs9\" (UniqueName: \"kubernetes.io/projected/b72e6503-3f54-43d9-ae0f-0240e1411010-kube-api-access-qrqs9\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.370930 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-utilities\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.371053 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-catalog-content\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.372257 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-catalog-content\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.372294 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-utilities\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.406011 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrqs9\" (UniqueName: \"kubernetes.io/projected/b72e6503-3f54-43d9-ae0f-0240e1411010-kube-api-access-qrqs9\") pod \"certified-operators-cjskg\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:36:59 crc kubenswrapper[4960]: I1002 08:36:59.495647 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:37:00 crc kubenswrapper[4960]: I1002 08:37:00.095939 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjskg"] Oct 02 08:37:00 crc kubenswrapper[4960]: I1002 08:37:00.241658 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjskg" event={"ID":"b72e6503-3f54-43d9-ae0f-0240e1411010","Type":"ContainerStarted","Data":"1616b75f772e7da9e8ceb7cc10a1baa96e1b5a168a8abc96bedcc3011000326e"} Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.250084 4960 generic.go:334] "Generic (PLEG): container finished" podID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerID="9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6" exitCode=0 Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.250218 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjskg" event={"ID":"b72e6503-3f54-43d9-ae0f-0240e1411010","Type":"ContainerDied","Data":"9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6"} Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.553416 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-78v5t"] Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.556590 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.575438 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-78v5t"] Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.634764 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-catalog-content\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.634828 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-utilities\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.634942 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph4tt\" (UniqueName: \"kubernetes.io/projected/f818fd68-45f6-4258-94c5-19a5546dc0b1-kube-api-access-ph4tt\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.736884 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph4tt\" (UniqueName: \"kubernetes.io/projected/f818fd68-45f6-4258-94c5-19a5546dc0b1-kube-api-access-ph4tt\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.737026 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-catalog-content\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.737071 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-utilities\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.737726 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-utilities\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.737884 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-catalog-content\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.762049 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph4tt\" (UniqueName: \"kubernetes.io/projected/f818fd68-45f6-4258-94c5-19a5546dc0b1-kube-api-access-ph4tt\") pod \"community-operators-78v5t\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:01 crc kubenswrapper[4960]: I1002 08:37:01.909769 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:02 crc kubenswrapper[4960]: I1002 08:37:02.269097 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjskg" event={"ID":"b72e6503-3f54-43d9-ae0f-0240e1411010","Type":"ContainerStarted","Data":"55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13"} Oct 02 08:37:02 crc kubenswrapper[4960]: I1002 08:37:02.484003 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-78v5t"] Oct 02 08:37:02 crc kubenswrapper[4960]: W1002 08:37:02.489224 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf818fd68_45f6_4258_94c5_19a5546dc0b1.slice/crio-f833c7c7641cc648de0289269969f2d368352b03d8bb4c31a75f522a735a61be WatchSource:0}: Error finding container f833c7c7641cc648de0289269969f2d368352b03d8bb4c31a75f522a735a61be: Status 404 returned error can't find the container with id f833c7c7641cc648de0289269969f2d368352b03d8bb4c31a75f522a735a61be Oct 02 08:37:03 crc kubenswrapper[4960]: I1002 08:37:03.285250 4960 generic.go:334] "Generic (PLEG): container finished" podID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerID="431081b1575ee17dccd3ec34a18e08cf2e71ba042be4153a983485429dc61cf8" exitCode=0 Oct 02 08:37:03 crc kubenswrapper[4960]: I1002 08:37:03.285335 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-78v5t" event={"ID":"f818fd68-45f6-4258-94c5-19a5546dc0b1","Type":"ContainerDied","Data":"431081b1575ee17dccd3ec34a18e08cf2e71ba042be4153a983485429dc61cf8"} Oct 02 08:37:03 crc kubenswrapper[4960]: I1002 08:37:03.285913 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-78v5t" event={"ID":"f818fd68-45f6-4258-94c5-19a5546dc0b1","Type":"ContainerStarted","Data":"f833c7c7641cc648de0289269969f2d368352b03d8bb4c31a75f522a735a61be"} Oct 02 08:37:04 crc kubenswrapper[4960]: I1002 08:37:04.297208 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-78v5t" event={"ID":"f818fd68-45f6-4258-94c5-19a5546dc0b1","Type":"ContainerStarted","Data":"3a24634b0227de9b04e92791e409ad732e311b84afdbe22b06010602607231c9"} Oct 02 08:37:04 crc kubenswrapper[4960]: I1002 08:37:04.301063 4960 generic.go:334] "Generic (PLEG): container finished" podID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerID="55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13" exitCode=0 Oct 02 08:37:04 crc kubenswrapper[4960]: I1002 08:37:04.301195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjskg" event={"ID":"b72e6503-3f54-43d9-ae0f-0240e1411010","Type":"ContainerDied","Data":"55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13"} Oct 02 08:37:06 crc kubenswrapper[4960]: I1002 08:37:06.333548 4960 generic.go:334] "Generic (PLEG): container finished" podID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerID="3a24634b0227de9b04e92791e409ad732e311b84afdbe22b06010602607231c9" exitCode=0 Oct 02 08:37:06 crc kubenswrapper[4960]: I1002 08:37:06.349283 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-78v5t" event={"ID":"f818fd68-45f6-4258-94c5-19a5546dc0b1","Type":"ContainerDied","Data":"3a24634b0227de9b04e92791e409ad732e311b84afdbe22b06010602607231c9"} Oct 02 08:37:06 crc kubenswrapper[4960]: I1002 08:37:06.349356 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjskg" event={"ID":"b72e6503-3f54-43d9-ae0f-0240e1411010","Type":"ContainerStarted","Data":"c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725"} Oct 02 08:37:06 crc kubenswrapper[4960]: I1002 08:37:06.382009 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cjskg" podStartSLOduration=3.388933304 podStartE2EDuration="7.381952086s" podCreationTimestamp="2025-10-02 08:36:59 +0000 UTC" firstStartedPulling="2025-10-02 08:37:01.252390461 +0000 UTC m=+4842.284336768" lastFinishedPulling="2025-10-02 08:37:05.245409263 +0000 UTC m=+4846.277355550" observedRunningTime="2025-10-02 08:37:06.372259796 +0000 UTC m=+4847.404206113" watchObservedRunningTime="2025-10-02 08:37:06.381952086 +0000 UTC m=+4847.413898403" Oct 02 08:37:07 crc kubenswrapper[4960]: I1002 08:37:07.362560 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-78v5t" event={"ID":"f818fd68-45f6-4258-94c5-19a5546dc0b1","Type":"ContainerStarted","Data":"1cae1e191ebbd61f8ee8a391ef7fbfd4fc09b206b9f61fae05d05669a7075eed"} Oct 02 08:37:07 crc kubenswrapper[4960]: I1002 08:37:07.390514 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-78v5t" podStartSLOduration=2.816884271 podStartE2EDuration="6.390483772s" podCreationTimestamp="2025-10-02 08:37:01 +0000 UTC" firstStartedPulling="2025-10-02 08:37:03.287808797 +0000 UTC m=+4844.319755084" lastFinishedPulling="2025-10-02 08:37:06.861408298 +0000 UTC m=+4847.893354585" observedRunningTime="2025-10-02 08:37:07.380370241 +0000 UTC m=+4848.412316528" watchObservedRunningTime="2025-10-02 08:37:07.390483772 +0000 UTC m=+4848.422430069" Oct 02 08:37:09 crc kubenswrapper[4960]: I1002 08:37:09.495901 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:37:09 crc kubenswrapper[4960]: I1002 08:37:09.496269 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:37:09 crc kubenswrapper[4960]: I1002 08:37:09.566635 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:37:10 crc kubenswrapper[4960]: I1002 08:37:10.454551 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:37:10 crc kubenswrapper[4960]: I1002 08:37:10.735046 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjskg"] Oct 02 08:37:11 crc kubenswrapper[4960]: I1002 08:37:11.910261 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:11 crc kubenswrapper[4960]: I1002 08:37:11.910386 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:12 crc kubenswrapper[4960]: I1002 08:37:12.004874 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:12 crc kubenswrapper[4960]: I1002 08:37:12.414787 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cjskg" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="registry-server" containerID="cri-o://c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725" gracePeriod=2 Oct 02 08:37:12 crc kubenswrapper[4960]: I1002 08:37:12.491101 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.155292 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.255011 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrqs9\" (UniqueName: \"kubernetes.io/projected/b72e6503-3f54-43d9-ae0f-0240e1411010-kube-api-access-qrqs9\") pod \"b72e6503-3f54-43d9-ae0f-0240e1411010\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.256064 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-catalog-content\") pod \"b72e6503-3f54-43d9-ae0f-0240e1411010\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.256165 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-utilities\") pod \"b72e6503-3f54-43d9-ae0f-0240e1411010\" (UID: \"b72e6503-3f54-43d9-ae0f-0240e1411010\") " Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.257101 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-utilities" (OuterVolumeSpecName: "utilities") pod "b72e6503-3f54-43d9-ae0f-0240e1411010" (UID: "b72e6503-3f54-43d9-ae0f-0240e1411010"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.264529 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b72e6503-3f54-43d9-ae0f-0240e1411010-kube-api-access-qrqs9" (OuterVolumeSpecName: "kube-api-access-qrqs9") pod "b72e6503-3f54-43d9-ae0f-0240e1411010" (UID: "b72e6503-3f54-43d9-ae0f-0240e1411010"). InnerVolumeSpecName "kube-api-access-qrqs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.342740 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b72e6503-3f54-43d9-ae0f-0240e1411010" (UID: "b72e6503-3f54-43d9-ae0f-0240e1411010"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.360318 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.360353 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrqs9\" (UniqueName: \"kubernetes.io/projected/b72e6503-3f54-43d9-ae0f-0240e1411010-kube-api-access-qrqs9\") on node \"crc\" DevicePath \"\"" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.360368 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e6503-3f54-43d9-ae0f-0240e1411010-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.427468 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjskg" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.427486 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjskg" event={"ID":"b72e6503-3f54-43d9-ae0f-0240e1411010","Type":"ContainerDied","Data":"c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725"} Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.428406 4960 scope.go:117] "RemoveContainer" containerID="c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.437180 4960 generic.go:334] "Generic (PLEG): container finished" podID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerID="c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725" exitCode=0 Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.437374 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjskg" event={"ID":"b72e6503-3f54-43d9-ae0f-0240e1411010","Type":"ContainerDied","Data":"1616b75f772e7da9e8ceb7cc10a1baa96e1b5a168a8abc96bedcc3011000326e"} Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.469147 4960 scope.go:117] "RemoveContainer" containerID="55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.478338 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjskg"] Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.489817 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cjskg"] Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.500738 4960 scope.go:117] "RemoveContainer" containerID="9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.536631 4960 scope.go:117] "RemoveContainer" containerID="c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725" Oct 02 08:37:13 crc kubenswrapper[4960]: E1002 08:37:13.537051 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725\": container with ID starting with c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725 not found: ID does not exist" containerID="c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.537096 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725"} err="failed to get container status \"c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725\": rpc error: code = NotFound desc = could not find container \"c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725\": container with ID starting with c243dfc61bcd7a4516416fc14a0156bbc6386958540706b8bfe4d9c39cad8725 not found: ID does not exist" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.537128 4960 scope.go:117] "RemoveContainer" containerID="55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13" Oct 02 08:37:13 crc kubenswrapper[4960]: E1002 08:37:13.537618 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13\": container with ID starting with 55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13 not found: ID does not exist" containerID="55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.537642 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13"} err="failed to get container status \"55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13\": rpc error: code = NotFound desc = could not find container \"55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13\": container with ID starting with 55e9b2469ac7e8ee5f4ff35196b4658d26048a5df7c9b611bbd851d431facb13 not found: ID does not exist" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.537658 4960 scope.go:117] "RemoveContainer" containerID="9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6" Oct 02 08:37:13 crc kubenswrapper[4960]: E1002 08:37:13.538118 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6\": container with ID starting with 9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6 not found: ID does not exist" containerID="9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6" Oct 02 08:37:13 crc kubenswrapper[4960]: I1002 08:37:13.538180 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6"} err="failed to get container status \"9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6\": rpc error: code = NotFound desc = could not find container \"9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6\": container with ID starting with 9e6567ce566a16d2801db631db222b94931d05b112575447c10f2b75fe778ed6 not found: ID does not exist" Oct 02 08:37:14 crc kubenswrapper[4960]: I1002 08:37:14.357807 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" path="/var/lib/kubelet/pods/b72e6503-3f54-43d9-ae0f-0240e1411010/volumes" Oct 02 08:37:14 crc kubenswrapper[4960]: I1002 08:37:14.359169 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-78v5t"] Oct 02 08:37:14 crc kubenswrapper[4960]: I1002 08:37:14.455385 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-78v5t" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="registry-server" containerID="cri-o://1cae1e191ebbd61f8ee8a391ef7fbfd4fc09b206b9f61fae05d05669a7075eed" gracePeriod=2 Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.466766 4960 generic.go:334] "Generic (PLEG): container finished" podID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerID="1cae1e191ebbd61f8ee8a391ef7fbfd4fc09b206b9f61fae05d05669a7075eed" exitCode=0 Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.466851 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-78v5t" event={"ID":"f818fd68-45f6-4258-94c5-19a5546dc0b1","Type":"ContainerDied","Data":"1cae1e191ebbd61f8ee8a391ef7fbfd4fc09b206b9f61fae05d05669a7075eed"} Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.618362 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.714367 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-utilities\") pod \"f818fd68-45f6-4258-94c5-19a5546dc0b1\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.714585 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph4tt\" (UniqueName: \"kubernetes.io/projected/f818fd68-45f6-4258-94c5-19a5546dc0b1-kube-api-access-ph4tt\") pod \"f818fd68-45f6-4258-94c5-19a5546dc0b1\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.714745 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-catalog-content\") pod \"f818fd68-45f6-4258-94c5-19a5546dc0b1\" (UID: \"f818fd68-45f6-4258-94c5-19a5546dc0b1\") " Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.715618 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-utilities" (OuterVolumeSpecName: "utilities") pod "f818fd68-45f6-4258-94c5-19a5546dc0b1" (UID: "f818fd68-45f6-4258-94c5-19a5546dc0b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.725363 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f818fd68-45f6-4258-94c5-19a5546dc0b1-kube-api-access-ph4tt" (OuterVolumeSpecName: "kube-api-access-ph4tt") pod "f818fd68-45f6-4258-94c5-19a5546dc0b1" (UID: "f818fd68-45f6-4258-94c5-19a5546dc0b1"). InnerVolumeSpecName "kube-api-access-ph4tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.768613 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f818fd68-45f6-4258-94c5-19a5546dc0b1" (UID: "f818fd68-45f6-4258-94c5-19a5546dc0b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.817457 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.817488 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f818fd68-45f6-4258-94c5-19a5546dc0b1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:37:15 crc kubenswrapper[4960]: I1002 08:37:15.817499 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph4tt\" (UniqueName: \"kubernetes.io/projected/f818fd68-45f6-4258-94c5-19a5546dc0b1-kube-api-access-ph4tt\") on node \"crc\" DevicePath \"\"" Oct 02 08:37:16 crc kubenswrapper[4960]: I1002 08:37:16.480923 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-78v5t" event={"ID":"f818fd68-45f6-4258-94c5-19a5546dc0b1","Type":"ContainerDied","Data":"f833c7c7641cc648de0289269969f2d368352b03d8bb4c31a75f522a735a61be"} Oct 02 08:37:16 crc kubenswrapper[4960]: I1002 08:37:16.481067 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-78v5t" Oct 02 08:37:16 crc kubenswrapper[4960]: I1002 08:37:16.481076 4960 scope.go:117] "RemoveContainer" containerID="1cae1e191ebbd61f8ee8a391ef7fbfd4fc09b206b9f61fae05d05669a7075eed" Oct 02 08:37:16 crc kubenswrapper[4960]: I1002 08:37:16.506043 4960 scope.go:117] "RemoveContainer" containerID="3a24634b0227de9b04e92791e409ad732e311b84afdbe22b06010602607231c9" Oct 02 08:37:16 crc kubenswrapper[4960]: I1002 08:37:16.526144 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-78v5t"] Oct 02 08:37:16 crc kubenswrapper[4960]: I1002 08:37:16.533396 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-78v5t"] Oct 02 08:37:16 crc kubenswrapper[4960]: I1002 08:37:16.538004 4960 scope.go:117] "RemoveContainer" containerID="431081b1575ee17dccd3ec34a18e08cf2e71ba042be4153a983485429dc61cf8" Oct 02 08:37:18 crc kubenswrapper[4960]: I1002 08:37:18.345314 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" path="/var/lib/kubelet/pods/f818fd68-45f6-4258-94c5-19a5546dc0b1/volumes" Oct 02 08:38:29 crc kubenswrapper[4960]: I1002 08:38:29.150918 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:38:29 crc kubenswrapper[4960]: I1002 08:38:29.151799 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:38:59 crc kubenswrapper[4960]: I1002 08:38:59.150454 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:38:59 crc kubenswrapper[4960]: I1002 08:38:59.151070 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.150327 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.151245 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.151313 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.152317 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.152376 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" gracePeriod=600 Oct 02 08:39:29 crc kubenswrapper[4960]: E1002 08:39:29.279497 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.906998 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" exitCode=0 Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.907075 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f"} Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.907144 4960 scope.go:117] "RemoveContainer" containerID="82652257eb990b346f43463872669f28c29971e7727a85fb74114ad3771937fe" Oct 02 08:39:29 crc kubenswrapper[4960]: I1002 08:39:29.907865 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:39:29 crc kubenswrapper[4960]: E1002 08:39:29.908168 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:39:42 crc kubenswrapper[4960]: I1002 08:39:42.331639 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:39:42 crc kubenswrapper[4960]: E1002 08:39:42.333090 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:39:57 crc kubenswrapper[4960]: I1002 08:39:57.329868 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:39:57 crc kubenswrapper[4960]: E1002 08:39:57.330902 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:40:09 crc kubenswrapper[4960]: I1002 08:40:09.331189 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:40:09 crc kubenswrapper[4960]: E1002 08:40:09.332380 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:40:20 crc kubenswrapper[4960]: I1002 08:40:20.340638 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:40:20 crc kubenswrapper[4960]: E1002 08:40:20.341598 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:40:34 crc kubenswrapper[4960]: I1002 08:40:34.330301 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:40:34 crc kubenswrapper[4960]: E1002 08:40:34.331226 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:40:46 crc kubenswrapper[4960]: I1002 08:40:46.330599 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:40:46 crc kubenswrapper[4960]: E1002 08:40:46.332380 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:40:57 crc kubenswrapper[4960]: I1002 08:40:57.330655 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:40:57 crc kubenswrapper[4960]: E1002 08:40:57.332100 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:41:12 crc kubenswrapper[4960]: I1002 08:41:12.331299 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:41:12 crc kubenswrapper[4960]: E1002 08:41:12.333336 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:41:27 crc kubenswrapper[4960]: I1002 08:41:27.331145 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:41:27 crc kubenswrapper[4960]: E1002 08:41:27.331917 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:41:40 crc kubenswrapper[4960]: I1002 08:41:40.358548 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:41:40 crc kubenswrapper[4960]: E1002 08:41:40.360136 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:41:52 crc kubenswrapper[4960]: I1002 08:41:52.330418 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:41:52 crc kubenswrapper[4960]: E1002 08:41:52.331394 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:42:06 crc kubenswrapper[4960]: I1002 08:42:06.331560 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:42:06 crc kubenswrapper[4960]: E1002 08:42:06.332376 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:42:18 crc kubenswrapper[4960]: I1002 08:42:18.331396 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:42:18 crc kubenswrapper[4960]: E1002 08:42:18.332284 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:42:30 crc kubenswrapper[4960]: I1002 08:42:30.338219 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:42:30 crc kubenswrapper[4960]: E1002 08:42:30.339060 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:42:41 crc kubenswrapper[4960]: I1002 08:42:41.330427 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:42:41 crc kubenswrapper[4960]: E1002 08:42:41.331343 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:42:54 crc kubenswrapper[4960]: I1002 08:42:54.330171 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:42:54 crc kubenswrapper[4960]: E1002 08:42:54.330831 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:43:09 crc kubenswrapper[4960]: I1002 08:43:09.331156 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:43:09 crc kubenswrapper[4960]: E1002 08:43:09.335089 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:43:21 crc kubenswrapper[4960]: I1002 08:43:21.329950 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:43:21 crc kubenswrapper[4960]: E1002 08:43:21.330765 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.264288 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pwbhd"] Oct 02 08:43:29 crc kubenswrapper[4960]: E1002 08:43:29.265711 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="registry-server" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.265727 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="registry-server" Oct 02 08:43:29 crc kubenswrapper[4960]: E1002 08:43:29.265748 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="registry-server" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.265753 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="registry-server" Oct 02 08:43:29 crc kubenswrapper[4960]: E1002 08:43:29.265775 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="extract-utilities" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.265781 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="extract-utilities" Oct 02 08:43:29 crc kubenswrapper[4960]: E1002 08:43:29.265794 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="extract-content" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.265800 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="extract-content" Oct 02 08:43:29 crc kubenswrapper[4960]: E1002 08:43:29.265808 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="extract-content" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.265813 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="extract-content" Oct 02 08:43:29 crc kubenswrapper[4960]: E1002 08:43:29.265832 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="extract-utilities" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.265838 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="extract-utilities" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.266078 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b72e6503-3f54-43d9-ae0f-0240e1411010" containerName="registry-server" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.266143 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f818fd68-45f6-4258-94c5-19a5546dc0b1" containerName="registry-server" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.267814 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.288948 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-catalog-content\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.289517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgn6h\" (UniqueName: \"kubernetes.io/projected/308b009c-0514-428a-a52f-15df74e15478-kube-api-access-fgn6h\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.289969 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-utilities\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.298372 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwbhd"] Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.392376 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-utilities\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.392500 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-catalog-content\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.392652 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgn6h\" (UniqueName: \"kubernetes.io/projected/308b009c-0514-428a-a52f-15df74e15478-kube-api-access-fgn6h\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.393603 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-utilities\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.393740 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-catalog-content\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.424276 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgn6h\" (UniqueName: \"kubernetes.io/projected/308b009c-0514-428a-a52f-15df74e15478-kube-api-access-fgn6h\") pod \"redhat-marketplace-pwbhd\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:29 crc kubenswrapper[4960]: I1002 08:43:29.598752 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:30 crc kubenswrapper[4960]: I1002 08:43:30.091282 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwbhd"] Oct 02 08:43:30 crc kubenswrapper[4960]: I1002 08:43:30.527865 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwbhd" event={"ID":"308b009c-0514-428a-a52f-15df74e15478","Type":"ContainerDied","Data":"65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863"} Oct 02 08:43:30 crc kubenswrapper[4960]: I1002 08:43:30.527819 4960 generic.go:334] "Generic (PLEG): container finished" podID="308b009c-0514-428a-a52f-15df74e15478" containerID="65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863" exitCode=0 Oct 02 08:43:30 crc kubenswrapper[4960]: I1002 08:43:30.530031 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwbhd" event={"ID":"308b009c-0514-428a-a52f-15df74e15478","Type":"ContainerStarted","Data":"ce09f298e2619d8d1901323759a2751cf83d0c724fa721ab7c8a69d93a80f9dc"} Oct 02 08:43:30 crc kubenswrapper[4960]: I1002 08:43:30.534883 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:43:31 crc kubenswrapper[4960]: I1002 08:43:31.544729 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwbhd" event={"ID":"308b009c-0514-428a-a52f-15df74e15478","Type":"ContainerStarted","Data":"49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7"} Oct 02 08:43:32 crc kubenswrapper[4960]: I1002 08:43:32.560663 4960 generic.go:334] "Generic (PLEG): container finished" podID="308b009c-0514-428a-a52f-15df74e15478" containerID="49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7" exitCode=0 Oct 02 08:43:32 crc kubenswrapper[4960]: I1002 08:43:32.560833 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwbhd" event={"ID":"308b009c-0514-428a-a52f-15df74e15478","Type":"ContainerDied","Data":"49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7"} Oct 02 08:43:33 crc kubenswrapper[4960]: I1002 08:43:33.577875 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwbhd" event={"ID":"308b009c-0514-428a-a52f-15df74e15478","Type":"ContainerStarted","Data":"2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8"} Oct 02 08:43:33 crc kubenswrapper[4960]: I1002 08:43:33.609561 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pwbhd" podStartSLOduration=2.144389539 podStartE2EDuration="4.609532729s" podCreationTimestamp="2025-10-02 08:43:29 +0000 UTC" firstStartedPulling="2025-10-02 08:43:30.534497886 +0000 UTC m=+5231.566444183" lastFinishedPulling="2025-10-02 08:43:32.999641066 +0000 UTC m=+5234.031587373" observedRunningTime="2025-10-02 08:43:33.600753921 +0000 UTC m=+5234.632700218" watchObservedRunningTime="2025-10-02 08:43:33.609532729 +0000 UTC m=+5234.641479026" Oct 02 08:43:35 crc kubenswrapper[4960]: I1002 08:43:35.330038 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:43:35 crc kubenswrapper[4960]: E1002 08:43:35.330712 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:43:39 crc kubenswrapper[4960]: I1002 08:43:39.599456 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:39 crc kubenswrapper[4960]: I1002 08:43:39.599875 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:39 crc kubenswrapper[4960]: I1002 08:43:39.688110 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:39 crc kubenswrapper[4960]: I1002 08:43:39.747411 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:39 crc kubenswrapper[4960]: I1002 08:43:39.930616 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwbhd"] Oct 02 08:43:41 crc kubenswrapper[4960]: I1002 08:43:41.702384 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pwbhd" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="registry-server" containerID="cri-o://2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8" gracePeriod=2 Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.451593 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.551106 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-catalog-content\") pod \"308b009c-0514-428a-a52f-15df74e15478\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.551250 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgn6h\" (UniqueName: \"kubernetes.io/projected/308b009c-0514-428a-a52f-15df74e15478-kube-api-access-fgn6h\") pod \"308b009c-0514-428a-a52f-15df74e15478\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.551408 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-utilities\") pod \"308b009c-0514-428a-a52f-15df74e15478\" (UID: \"308b009c-0514-428a-a52f-15df74e15478\") " Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.552504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-utilities" (OuterVolumeSpecName: "utilities") pod "308b009c-0514-428a-a52f-15df74e15478" (UID: "308b009c-0514-428a-a52f-15df74e15478"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.559222 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308b009c-0514-428a-a52f-15df74e15478-kube-api-access-fgn6h" (OuterVolumeSpecName: "kube-api-access-fgn6h") pod "308b009c-0514-428a-a52f-15df74e15478" (UID: "308b009c-0514-428a-a52f-15df74e15478"). InnerVolumeSpecName "kube-api-access-fgn6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.564331 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "308b009c-0514-428a-a52f-15df74e15478" (UID: "308b009c-0514-428a-a52f-15df74e15478"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.653786 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.653837 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308b009c-0514-428a-a52f-15df74e15478-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.653852 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgn6h\" (UniqueName: \"kubernetes.io/projected/308b009c-0514-428a-a52f-15df74e15478-kube-api-access-fgn6h\") on node \"crc\" DevicePath \"\"" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.713208 4960 generic.go:334] "Generic (PLEG): container finished" podID="308b009c-0514-428a-a52f-15df74e15478" containerID="2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8" exitCode=0 Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.713276 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwbhd" event={"ID":"308b009c-0514-428a-a52f-15df74e15478","Type":"ContainerDied","Data":"2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8"} Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.713363 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwbhd" event={"ID":"308b009c-0514-428a-a52f-15df74e15478","Type":"ContainerDied","Data":"ce09f298e2619d8d1901323759a2751cf83d0c724fa721ab7c8a69d93a80f9dc"} Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.713390 4960 scope.go:117] "RemoveContainer" containerID="2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.714844 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwbhd" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.738651 4960 scope.go:117] "RemoveContainer" containerID="49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.767672 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwbhd"] Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.779593 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwbhd"] Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.782863 4960 scope.go:117] "RemoveContainer" containerID="65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.818297 4960 scope.go:117] "RemoveContainer" containerID="2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8" Oct 02 08:43:42 crc kubenswrapper[4960]: E1002 08:43:42.818899 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8\": container with ID starting with 2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8 not found: ID does not exist" containerID="2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.818962 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8"} err="failed to get container status \"2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8\": rpc error: code = NotFound desc = could not find container \"2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8\": container with ID starting with 2f517e4065dfaab2d01e707d6a1c48579f42c4d9675b4bd9e0d64798aa512cf8 not found: ID does not exist" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.819016 4960 scope.go:117] "RemoveContainer" containerID="49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7" Oct 02 08:43:42 crc kubenswrapper[4960]: E1002 08:43:42.822766 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7\": container with ID starting with 49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7 not found: ID does not exist" containerID="49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.822905 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7"} err="failed to get container status \"49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7\": rpc error: code = NotFound desc = could not find container \"49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7\": container with ID starting with 49e342ac903d7a9f8d745e06cb6bbd6a93ecd00080682f381f138cf0ec16f8a7 not found: ID does not exist" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.823023 4960 scope.go:117] "RemoveContainer" containerID="65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863" Oct 02 08:43:42 crc kubenswrapper[4960]: E1002 08:43:42.823735 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863\": container with ID starting with 65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863 not found: ID does not exist" containerID="65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863" Oct 02 08:43:42 crc kubenswrapper[4960]: I1002 08:43:42.823776 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863"} err="failed to get container status \"65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863\": rpc error: code = NotFound desc = could not find container \"65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863\": container with ID starting with 65745b1a9ea3e7cb4b80c4ce07cfb9ab7bdeaacfc276f250021ef93d3d38c863 not found: ID does not exist" Oct 02 08:43:44 crc kubenswrapper[4960]: I1002 08:43:44.345662 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308b009c-0514-428a-a52f-15df74e15478" path="/var/lib/kubelet/pods/308b009c-0514-428a-a52f-15df74e15478/volumes" Oct 02 08:43:47 crc kubenswrapper[4960]: I1002 08:43:47.330400 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:43:47 crc kubenswrapper[4960]: E1002 08:43:47.331710 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:44:00 crc kubenswrapper[4960]: I1002 08:44:00.344274 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:44:00 crc kubenswrapper[4960]: E1002 08:44:00.345357 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:44:11 crc kubenswrapper[4960]: I1002 08:44:11.330901 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:44:11 crc kubenswrapper[4960]: E1002 08:44:11.332675 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:44:25 crc kubenswrapper[4960]: I1002 08:44:25.329754 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:44:25 crc kubenswrapper[4960]: E1002 08:44:25.330663 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:44:40 crc kubenswrapper[4960]: I1002 08:44:40.337903 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:44:41 crc kubenswrapper[4960]: I1002 08:44:41.394287 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"0c87b397fcffb54747129e24e768259dab6847d138d75073f9845299de9429a7"} Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.150354 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x"] Oct 02 08:45:00 crc kubenswrapper[4960]: E1002 08:45:00.151858 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="extract-utilities" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.151878 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="extract-utilities" Oct 02 08:45:00 crc kubenswrapper[4960]: E1002 08:45:00.151939 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="extract-content" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.151950 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="extract-content" Oct 02 08:45:00 crc kubenswrapper[4960]: E1002 08:45:00.151968 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="registry-server" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.151977 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="registry-server" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.152263 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="308b009c-0514-428a-a52f-15df74e15478" containerName="registry-server" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.153195 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.155883 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.157123 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.165948 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x"] Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.228838 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-config-volume\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.229363 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-secret-volume\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.230119 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fvqx\" (UniqueName: \"kubernetes.io/projected/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-kube-api-access-5fvqx\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.332247 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fvqx\" (UniqueName: \"kubernetes.io/projected/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-kube-api-access-5fvqx\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.332333 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-config-volume\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.332400 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-secret-volume\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.334597 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-config-volume\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.341688 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-secret-volume\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.367872 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fvqx\" (UniqueName: \"kubernetes.io/projected/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-kube-api-access-5fvqx\") pod \"collect-profiles-29323245-5495x\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.490680 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:00 crc kubenswrapper[4960]: I1002 08:45:00.998384 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x"] Oct 02 08:45:00 crc kubenswrapper[4960]: W1002 08:45:00.999265 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd64343f3_cd2f_4722_a9d9_2d12360dc6f5.slice/crio-933a3d1a05c6bea596007f086f55d0fa93049ab56ccafd46941b2ef3231f28a4 WatchSource:0}: Error finding container 933a3d1a05c6bea596007f086f55d0fa93049ab56ccafd46941b2ef3231f28a4: Status 404 returned error can't find the container with id 933a3d1a05c6bea596007f086f55d0fa93049ab56ccafd46941b2ef3231f28a4 Oct 02 08:45:01 crc kubenswrapper[4960]: I1002 08:45:01.593849 4960 generic.go:334] "Generic (PLEG): container finished" podID="d64343f3-cd2f-4722-a9d9-2d12360dc6f5" containerID="20962b7aff238872d3935749a500fcf85ace57b25018940be68c9777345c0000" exitCode=0 Oct 02 08:45:01 crc kubenswrapper[4960]: I1002 08:45:01.594195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" event={"ID":"d64343f3-cd2f-4722-a9d9-2d12360dc6f5","Type":"ContainerDied","Data":"20962b7aff238872d3935749a500fcf85ace57b25018940be68c9777345c0000"} Oct 02 08:45:01 crc kubenswrapper[4960]: I1002 08:45:01.594401 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" event={"ID":"d64343f3-cd2f-4722-a9d9-2d12360dc6f5","Type":"ContainerStarted","Data":"933a3d1a05c6bea596007f086f55d0fa93049ab56ccafd46941b2ef3231f28a4"} Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.137126 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.310125 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-config-volume\") pod \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.310515 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fvqx\" (UniqueName: \"kubernetes.io/projected/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-kube-api-access-5fvqx\") pod \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.310578 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-secret-volume\") pod \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\" (UID: \"d64343f3-cd2f-4722-a9d9-2d12360dc6f5\") " Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.311180 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-config-volume" (OuterVolumeSpecName: "config-volume") pod "d64343f3-cd2f-4722-a9d9-2d12360dc6f5" (UID: "d64343f3-cd2f-4722-a9d9-2d12360dc6f5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.319065 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d64343f3-cd2f-4722-a9d9-2d12360dc6f5" (UID: "d64343f3-cd2f-4722-a9d9-2d12360dc6f5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.320669 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-kube-api-access-5fvqx" (OuterVolumeSpecName: "kube-api-access-5fvqx") pod "d64343f3-cd2f-4722-a9d9-2d12360dc6f5" (UID: "d64343f3-cd2f-4722-a9d9-2d12360dc6f5"). InnerVolumeSpecName "kube-api-access-5fvqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.416607 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fvqx\" (UniqueName: \"kubernetes.io/projected/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-kube-api-access-5fvqx\") on node \"crc\" DevicePath \"\"" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.417050 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.417065 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d64343f3-cd2f-4722-a9d9-2d12360dc6f5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.614162 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" event={"ID":"d64343f3-cd2f-4722-a9d9-2d12360dc6f5","Type":"ContainerDied","Data":"933a3d1a05c6bea596007f086f55d0fa93049ab56ccafd46941b2ef3231f28a4"} Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.614207 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="933a3d1a05c6bea596007f086f55d0fa93049ab56ccafd46941b2ef3231f28a4" Oct 02 08:45:03 crc kubenswrapper[4960]: I1002 08:45:03.614275 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x" Oct 02 08:45:04 crc kubenswrapper[4960]: I1002 08:45:04.240917 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz"] Oct 02 08:45:04 crc kubenswrapper[4960]: I1002 08:45:04.253460 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-ppgrz"] Oct 02 08:45:04 crc kubenswrapper[4960]: I1002 08:45:04.348647 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d666c8-3d64-4297-89eb-07c7e0a316e6" path="/var/lib/kubelet/pods/f5d666c8-3d64-4297-89eb-07c7e0a316e6/volumes" Oct 02 08:46:03 crc kubenswrapper[4960]: I1002 08:46:03.323824 4960 scope.go:117] "RemoveContainer" containerID="14041f654a74645cca51e60a458eaef5fa73bbf411b3e005c22100b4e66c89dc" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.424718 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-br55k"] Oct 02 08:46:46 crc kubenswrapper[4960]: E1002 08:46:46.428186 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64343f3-cd2f-4722-a9d9-2d12360dc6f5" containerName="collect-profiles" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.428278 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64343f3-cd2f-4722-a9d9-2d12360dc6f5" containerName="collect-profiles" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.428527 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64343f3-cd2f-4722-a9d9-2d12360dc6f5" containerName="collect-profiles" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.430457 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.434455 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-br55k"] Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.609202 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-utilities\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.609539 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-catalog-content\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.610658 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m8x8\" (UniqueName: \"kubernetes.io/projected/5964d589-4c47-4cf0-8230-461050e4b767-kube-api-access-4m8x8\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.713357 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-utilities\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.713522 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-catalog-content\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.713600 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m8x8\" (UniqueName: \"kubernetes.io/projected/5964d589-4c47-4cf0-8230-461050e4b767-kube-api-access-4m8x8\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.713967 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-utilities\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.714093 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-catalog-content\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.746038 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m8x8\" (UniqueName: \"kubernetes.io/projected/5964d589-4c47-4cf0-8230-461050e4b767-kube-api-access-4m8x8\") pod \"redhat-operators-br55k\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:46 crc kubenswrapper[4960]: I1002 08:46:46.750818 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:47 crc kubenswrapper[4960]: I1002 08:46:47.330902 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-br55k"] Oct 02 08:46:47 crc kubenswrapper[4960]: I1002 08:46:47.669773 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br55k" event={"ID":"5964d589-4c47-4cf0-8230-461050e4b767","Type":"ContainerStarted","Data":"3d99ec3689dcd23e5d60d0ad084b485d9e3e2554dfbf2a89bf3858c997b41b68"} Oct 02 08:46:48 crc kubenswrapper[4960]: I1002 08:46:48.680495 4960 generic.go:334] "Generic (PLEG): container finished" podID="5964d589-4c47-4cf0-8230-461050e4b767" containerID="271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55" exitCode=0 Oct 02 08:46:48 crc kubenswrapper[4960]: I1002 08:46:48.680542 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br55k" event={"ID":"5964d589-4c47-4cf0-8230-461050e4b767","Type":"ContainerDied","Data":"271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55"} Oct 02 08:46:50 crc kubenswrapper[4960]: I1002 08:46:50.699737 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br55k" event={"ID":"5964d589-4c47-4cf0-8230-461050e4b767","Type":"ContainerStarted","Data":"33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7"} Oct 02 08:46:53 crc kubenswrapper[4960]: I1002 08:46:53.734424 4960 generic.go:334] "Generic (PLEG): container finished" podID="5964d589-4c47-4cf0-8230-461050e4b767" containerID="33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7" exitCode=0 Oct 02 08:46:53 crc kubenswrapper[4960]: I1002 08:46:53.734534 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br55k" event={"ID":"5964d589-4c47-4cf0-8230-461050e4b767","Type":"ContainerDied","Data":"33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7"} Oct 02 08:46:54 crc kubenswrapper[4960]: I1002 08:46:54.750640 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br55k" event={"ID":"5964d589-4c47-4cf0-8230-461050e4b767","Type":"ContainerStarted","Data":"c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927"} Oct 02 08:46:54 crc kubenswrapper[4960]: I1002 08:46:54.775329 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-br55k" podStartSLOduration=3.038587809 podStartE2EDuration="8.775299608s" podCreationTimestamp="2025-10-02 08:46:46 +0000 UTC" firstStartedPulling="2025-10-02 08:46:48.683042196 +0000 UTC m=+5429.714988503" lastFinishedPulling="2025-10-02 08:46:54.419754015 +0000 UTC m=+5435.451700302" observedRunningTime="2025-10-02 08:46:54.77424769 +0000 UTC m=+5435.806194037" watchObservedRunningTime="2025-10-02 08:46:54.775299608 +0000 UTC m=+5435.807245915" Oct 02 08:46:56 crc kubenswrapper[4960]: I1002 08:46:56.751945 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:56 crc kubenswrapper[4960]: I1002 08:46:56.752658 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:46:57 crc kubenswrapper[4960]: I1002 08:46:57.806115 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-br55k" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="registry-server" probeResult="failure" output=< Oct 02 08:46:57 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 08:46:57 crc kubenswrapper[4960]: > Oct 02 08:46:59 crc kubenswrapper[4960]: I1002 08:46:59.150341 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:46:59 crc kubenswrapper[4960]: I1002 08:46:59.150694 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:47:06 crc kubenswrapper[4960]: I1002 08:47:06.792543 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:47:06 crc kubenswrapper[4960]: I1002 08:47:06.845671 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:47:07 crc kubenswrapper[4960]: I1002 08:47:07.031565 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-br55k"] Oct 02 08:47:07 crc kubenswrapper[4960]: I1002 08:47:07.895526 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-br55k" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="registry-server" containerID="cri-o://c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927" gracePeriod=2 Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.495850 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.593826 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-catalog-content\") pod \"5964d589-4c47-4cf0-8230-461050e4b767\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.594057 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-utilities\") pod \"5964d589-4c47-4cf0-8230-461050e4b767\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.594340 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m8x8\" (UniqueName: \"kubernetes.io/projected/5964d589-4c47-4cf0-8230-461050e4b767-kube-api-access-4m8x8\") pod \"5964d589-4c47-4cf0-8230-461050e4b767\" (UID: \"5964d589-4c47-4cf0-8230-461050e4b767\") " Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.595666 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-utilities" (OuterVolumeSpecName: "utilities") pod "5964d589-4c47-4cf0-8230-461050e4b767" (UID: "5964d589-4c47-4cf0-8230-461050e4b767"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.607370 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5964d589-4c47-4cf0-8230-461050e4b767-kube-api-access-4m8x8" (OuterVolumeSpecName: "kube-api-access-4m8x8") pod "5964d589-4c47-4cf0-8230-461050e4b767" (UID: "5964d589-4c47-4cf0-8230-461050e4b767"). InnerVolumeSpecName "kube-api-access-4m8x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.673663 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5964d589-4c47-4cf0-8230-461050e4b767" (UID: "5964d589-4c47-4cf0-8230-461050e4b767"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.697728 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.697769 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5964d589-4c47-4cf0-8230-461050e4b767-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.697780 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m8x8\" (UniqueName: \"kubernetes.io/projected/5964d589-4c47-4cf0-8230-461050e4b767-kube-api-access-4m8x8\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.914406 4960 generic.go:334] "Generic (PLEG): container finished" podID="5964d589-4c47-4cf0-8230-461050e4b767" containerID="c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927" exitCode=0 Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.914465 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br55k" event={"ID":"5964d589-4c47-4cf0-8230-461050e4b767","Type":"ContainerDied","Data":"c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927"} Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.914507 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br55k" event={"ID":"5964d589-4c47-4cf0-8230-461050e4b767","Type":"ContainerDied","Data":"3d99ec3689dcd23e5d60d0ad084b485d9e3e2554dfbf2a89bf3858c997b41b68"} Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.914529 4960 scope.go:117] "RemoveContainer" containerID="c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.914731 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br55k" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.989709 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-br55k"] Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.991861 4960 scope.go:117] "RemoveContainer" containerID="33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7" Oct 02 08:47:08 crc kubenswrapper[4960]: I1002 08:47:08.999370 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-br55k"] Oct 02 08:47:09 crc kubenswrapper[4960]: I1002 08:47:09.023869 4960 scope.go:117] "RemoveContainer" containerID="271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55" Oct 02 08:47:09 crc kubenswrapper[4960]: I1002 08:47:09.066863 4960 scope.go:117] "RemoveContainer" containerID="c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927" Oct 02 08:47:09 crc kubenswrapper[4960]: E1002 08:47:09.067958 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927\": container with ID starting with c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927 not found: ID does not exist" containerID="c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927" Oct 02 08:47:09 crc kubenswrapper[4960]: I1002 08:47:09.068073 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927"} err="failed to get container status \"c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927\": rpc error: code = NotFound desc = could not find container \"c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927\": container with ID starting with c3f2ebc97b056aaa601380e32634b42a559c95fc0d44d81e039d985fb50c8927 not found: ID does not exist" Oct 02 08:47:09 crc kubenswrapper[4960]: I1002 08:47:09.068106 4960 scope.go:117] "RemoveContainer" containerID="33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7" Oct 02 08:47:09 crc kubenswrapper[4960]: E1002 08:47:09.068584 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7\": container with ID starting with 33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7 not found: ID does not exist" containerID="33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7" Oct 02 08:47:09 crc kubenswrapper[4960]: I1002 08:47:09.068608 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7"} err="failed to get container status \"33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7\": rpc error: code = NotFound desc = could not find container \"33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7\": container with ID starting with 33782cbf4d5a41722d932c4988a4dc2c9a754345612c79fb12f7429265453ed7 not found: ID does not exist" Oct 02 08:47:09 crc kubenswrapper[4960]: I1002 08:47:09.068621 4960 scope.go:117] "RemoveContainer" containerID="271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55" Oct 02 08:47:09 crc kubenswrapper[4960]: E1002 08:47:09.069007 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55\": container with ID starting with 271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55 not found: ID does not exist" containerID="271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55" Oct 02 08:47:09 crc kubenswrapper[4960]: I1002 08:47:09.069035 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55"} err="failed to get container status \"271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55\": rpc error: code = NotFound desc = could not find container \"271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55\": container with ID starting with 271e9aa1ae59bd29c86e0b6a10fcd3325afda76c1b22b7d90bed1f914d7e5a55 not found: ID does not exist" Oct 02 08:47:10 crc kubenswrapper[4960]: I1002 08:47:10.344747 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5964d589-4c47-4cf0-8230-461050e4b767" path="/var/lib/kubelet/pods/5964d589-4c47-4cf0-8230-461050e4b767/volumes" Oct 02 08:47:29 crc kubenswrapper[4960]: I1002 08:47:29.150721 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:47:29 crc kubenswrapper[4960]: I1002 08:47:29.151464 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.006627 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s67vh"] Oct 02 08:47:36 crc kubenswrapper[4960]: E1002 08:47:36.008061 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="registry-server" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.008082 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="registry-server" Oct 02 08:47:36 crc kubenswrapper[4960]: E1002 08:47:36.008101 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="extract-content" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.008109 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="extract-content" Oct 02 08:47:36 crc kubenswrapper[4960]: E1002 08:47:36.008127 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="extract-utilities" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.008136 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="extract-utilities" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.008442 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5964d589-4c47-4cf0-8230-461050e4b767" containerName="registry-server" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.010670 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.021918 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s67vh"] Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.135443 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2gp7\" (UniqueName: \"kubernetes.io/projected/89bf13aa-d944-4ffc-9efc-e8998765a417-kube-api-access-g2gp7\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.135522 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-catalog-content\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.135929 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-utilities\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.238462 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-catalog-content\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.238602 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-utilities\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.238760 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2gp7\" (UniqueName: \"kubernetes.io/projected/89bf13aa-d944-4ffc-9efc-e8998765a417-kube-api-access-g2gp7\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.239101 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-catalog-content\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.239215 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-utilities\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.262190 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2gp7\" (UniqueName: \"kubernetes.io/projected/89bf13aa-d944-4ffc-9efc-e8998765a417-kube-api-access-g2gp7\") pod \"certified-operators-s67vh\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.357096 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:36 crc kubenswrapper[4960]: I1002 08:47:36.940500 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s67vh"] Oct 02 08:47:38 crc kubenswrapper[4960]: I1002 08:47:38.226425 4960 generic.go:334] "Generic (PLEG): container finished" podID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerID="d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8" exitCode=0 Oct 02 08:47:38 crc kubenswrapper[4960]: I1002 08:47:38.226576 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s67vh" event={"ID":"89bf13aa-d944-4ffc-9efc-e8998765a417","Type":"ContainerDied","Data":"d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8"} Oct 02 08:47:38 crc kubenswrapper[4960]: I1002 08:47:38.227266 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s67vh" event={"ID":"89bf13aa-d944-4ffc-9efc-e8998765a417","Type":"ContainerStarted","Data":"6038cd26f0c2cbf08545e9aa7d06359100537fd980d8c4f6c5a2f107c9ce4008"} Oct 02 08:47:39 crc kubenswrapper[4960]: I1002 08:47:39.242014 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s67vh" event={"ID":"89bf13aa-d944-4ffc-9efc-e8998765a417","Type":"ContainerStarted","Data":"3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe"} Oct 02 08:47:41 crc kubenswrapper[4960]: I1002 08:47:41.271788 4960 generic.go:334] "Generic (PLEG): container finished" podID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerID="3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe" exitCode=0 Oct 02 08:47:41 crc kubenswrapper[4960]: I1002 08:47:41.271887 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s67vh" event={"ID":"89bf13aa-d944-4ffc-9efc-e8998765a417","Type":"ContainerDied","Data":"3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe"} Oct 02 08:47:42 crc kubenswrapper[4960]: I1002 08:47:42.290303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s67vh" event={"ID":"89bf13aa-d944-4ffc-9efc-e8998765a417","Type":"ContainerStarted","Data":"dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f"} Oct 02 08:47:42 crc kubenswrapper[4960]: I1002 08:47:42.326994 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s67vh" podStartSLOduration=3.753766892 podStartE2EDuration="7.326951422s" podCreationTimestamp="2025-10-02 08:47:35 +0000 UTC" firstStartedPulling="2025-10-02 08:47:38.230631689 +0000 UTC m=+5479.262578016" lastFinishedPulling="2025-10-02 08:47:41.803816259 +0000 UTC m=+5482.835762546" observedRunningTime="2025-10-02 08:47:42.314898476 +0000 UTC m=+5483.346844793" watchObservedRunningTime="2025-10-02 08:47:42.326951422 +0000 UTC m=+5483.358897709" Oct 02 08:47:46 crc kubenswrapper[4960]: I1002 08:47:46.357922 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:46 crc kubenswrapper[4960]: I1002 08:47:46.358771 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:46 crc kubenswrapper[4960]: I1002 08:47:46.418060 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:47 crc kubenswrapper[4960]: I1002 08:47:47.426734 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:47 crc kubenswrapper[4960]: I1002 08:47:47.502851 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s67vh"] Oct 02 08:47:49 crc kubenswrapper[4960]: I1002 08:47:49.379330 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s67vh" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="registry-server" containerID="cri-o://dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f" gracePeriod=2 Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.070195 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.204971 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2gp7\" (UniqueName: \"kubernetes.io/projected/89bf13aa-d944-4ffc-9efc-e8998765a417-kube-api-access-g2gp7\") pod \"89bf13aa-d944-4ffc-9efc-e8998765a417\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.205088 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-utilities\") pod \"89bf13aa-d944-4ffc-9efc-e8998765a417\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.205263 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-catalog-content\") pod \"89bf13aa-d944-4ffc-9efc-e8998765a417\" (UID: \"89bf13aa-d944-4ffc-9efc-e8998765a417\") " Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.206198 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-utilities" (OuterVolumeSpecName: "utilities") pod "89bf13aa-d944-4ffc-9efc-e8998765a417" (UID: "89bf13aa-d944-4ffc-9efc-e8998765a417"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.214669 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89bf13aa-d944-4ffc-9efc-e8998765a417-kube-api-access-g2gp7" (OuterVolumeSpecName: "kube-api-access-g2gp7") pod "89bf13aa-d944-4ffc-9efc-e8998765a417" (UID: "89bf13aa-d944-4ffc-9efc-e8998765a417"). InnerVolumeSpecName "kube-api-access-g2gp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.267298 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89bf13aa-d944-4ffc-9efc-e8998765a417" (UID: "89bf13aa-d944-4ffc-9efc-e8998765a417"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.308217 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.308516 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2gp7\" (UniqueName: \"kubernetes.io/projected/89bf13aa-d944-4ffc-9efc-e8998765a417-kube-api-access-g2gp7\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.308584 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf13aa-d944-4ffc-9efc-e8998765a417-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.390413 4960 generic.go:334] "Generic (PLEG): container finished" podID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerID="dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f" exitCode=0 Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.390474 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s67vh" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.390497 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s67vh" event={"ID":"89bf13aa-d944-4ffc-9efc-e8998765a417","Type":"ContainerDied","Data":"dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f"} Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.392140 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s67vh" event={"ID":"89bf13aa-d944-4ffc-9efc-e8998765a417","Type":"ContainerDied","Data":"6038cd26f0c2cbf08545e9aa7d06359100537fd980d8c4f6c5a2f107c9ce4008"} Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.392173 4960 scope.go:117] "RemoveContainer" containerID="dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.425882 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s67vh"] Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.432251 4960 scope.go:117] "RemoveContainer" containerID="3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.436929 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s67vh"] Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.473530 4960 scope.go:117] "RemoveContainer" containerID="d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.539076 4960 scope.go:117] "RemoveContainer" containerID="dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f" Oct 02 08:47:50 crc kubenswrapper[4960]: E1002 08:47:50.540049 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f\": container with ID starting with dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f not found: ID does not exist" containerID="dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.540118 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f"} err="failed to get container status \"dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f\": rpc error: code = NotFound desc = could not find container \"dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f\": container with ID starting with dcf6729ce3426b728a8c3b1d5239f6bf5a1394cd94b6128b382dc825861f376f not found: ID does not exist" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.540157 4960 scope.go:117] "RemoveContainer" containerID="3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe" Oct 02 08:47:50 crc kubenswrapper[4960]: E1002 08:47:50.540675 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe\": container with ID starting with 3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe not found: ID does not exist" containerID="3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.540833 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe"} err="failed to get container status \"3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe\": rpc error: code = NotFound desc = could not find container \"3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe\": container with ID starting with 3eb04431315d7efa4bad034f2d72a482a7c765878233c888cdc66b52fcab73fe not found: ID does not exist" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.540920 4960 scope.go:117] "RemoveContainer" containerID="d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8" Oct 02 08:47:50 crc kubenswrapper[4960]: E1002 08:47:50.541610 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8\": container with ID starting with d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8 not found: ID does not exist" containerID="d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8" Oct 02 08:47:50 crc kubenswrapper[4960]: I1002 08:47:50.541664 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8"} err="failed to get container status \"d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8\": rpc error: code = NotFound desc = could not find container \"d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8\": container with ID starting with d19f2e0b35e369ac82214c21a9277807fd83a4f0d9d0d099a9f22ecad7e8b2a8 not found: ID does not exist" Oct 02 08:47:52 crc kubenswrapper[4960]: I1002 08:47:52.342109 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" path="/var/lib/kubelet/pods/89bf13aa-d944-4ffc-9efc-e8998765a417/volumes" Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.150220 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.151058 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.151123 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.152082 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c87b397fcffb54747129e24e768259dab6847d138d75073f9845299de9429a7"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.152131 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://0c87b397fcffb54747129e24e768259dab6847d138d75073f9845299de9429a7" gracePeriod=600 Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.537069 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="0c87b397fcffb54747129e24e768259dab6847d138d75073f9845299de9429a7" exitCode=0 Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.537122 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"0c87b397fcffb54747129e24e768259dab6847d138d75073f9845299de9429a7"} Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.537575 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49"} Oct 02 08:47:59 crc kubenswrapper[4960]: I1002 08:47:59.537608 4960 scope.go:117] "RemoveContainer" containerID="15b9aa4e0b026c70d33f2fdd89a67b40ec1d8d2a4d4863e8667f66a2d7c9011f" Oct 02 08:48:22 crc kubenswrapper[4960]: I1002 08:48:22.973862 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-znb9z"] Oct 02 08:48:22 crc kubenswrapper[4960]: E1002 08:48:22.974900 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="extract-utilities" Oct 02 08:48:22 crc kubenswrapper[4960]: I1002 08:48:22.974914 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="extract-utilities" Oct 02 08:48:22 crc kubenswrapper[4960]: E1002 08:48:22.974927 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="registry-server" Oct 02 08:48:22 crc kubenswrapper[4960]: I1002 08:48:22.974932 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="registry-server" Oct 02 08:48:22 crc kubenswrapper[4960]: E1002 08:48:22.974956 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="extract-content" Oct 02 08:48:22 crc kubenswrapper[4960]: I1002 08:48:22.974963 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="extract-content" Oct 02 08:48:22 crc kubenswrapper[4960]: I1002 08:48:22.975172 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="89bf13aa-d944-4ffc-9efc-e8998765a417" containerName="registry-server" Oct 02 08:48:22 crc kubenswrapper[4960]: I1002 08:48:22.976508 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:22 crc kubenswrapper[4960]: I1002 08:48:22.994262 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-znb9z"] Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.130424 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-utilities\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.130494 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-catalog-content\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.130588 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfglr\" (UniqueName: \"kubernetes.io/projected/b33b901e-2d7b-4f89-91b2-44615123ea7f-kube-api-access-kfglr\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.232384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-utilities\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.232429 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-catalog-content\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.232493 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfglr\" (UniqueName: \"kubernetes.io/projected/b33b901e-2d7b-4f89-91b2-44615123ea7f-kube-api-access-kfglr\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.233083 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-catalog-content\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.233279 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-utilities\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.260221 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfglr\" (UniqueName: \"kubernetes.io/projected/b33b901e-2d7b-4f89-91b2-44615123ea7f-kube-api-access-kfglr\") pod \"community-operators-znb9z\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.299934 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:23 crc kubenswrapper[4960]: I1002 08:48:23.866168 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-znb9z"] Oct 02 08:48:23 crc kubenswrapper[4960]: W1002 08:48:23.871499 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb33b901e_2d7b_4f89_91b2_44615123ea7f.slice/crio-e14a53d4e83a76065d0fc5b68171a0b4f2504050d9003f86815ce43f6de001b7 WatchSource:0}: Error finding container e14a53d4e83a76065d0fc5b68171a0b4f2504050d9003f86815ce43f6de001b7: Status 404 returned error can't find the container with id e14a53d4e83a76065d0fc5b68171a0b4f2504050d9003f86815ce43f6de001b7 Oct 02 08:48:24 crc kubenswrapper[4960]: I1002 08:48:24.791874 4960 generic.go:334] "Generic (PLEG): container finished" podID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerID="f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd" exitCode=0 Oct 02 08:48:24 crc kubenswrapper[4960]: I1002 08:48:24.792007 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znb9z" event={"ID":"b33b901e-2d7b-4f89-91b2-44615123ea7f","Type":"ContainerDied","Data":"f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd"} Oct 02 08:48:24 crc kubenswrapper[4960]: I1002 08:48:24.792441 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znb9z" event={"ID":"b33b901e-2d7b-4f89-91b2-44615123ea7f","Type":"ContainerStarted","Data":"e14a53d4e83a76065d0fc5b68171a0b4f2504050d9003f86815ce43f6de001b7"} Oct 02 08:48:25 crc kubenswrapper[4960]: I1002 08:48:25.802670 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znb9z" event={"ID":"b33b901e-2d7b-4f89-91b2-44615123ea7f","Type":"ContainerStarted","Data":"021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8"} Oct 02 08:48:27 crc kubenswrapper[4960]: I1002 08:48:27.821555 4960 generic.go:334] "Generic (PLEG): container finished" podID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerID="021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8" exitCode=0 Oct 02 08:48:27 crc kubenswrapper[4960]: I1002 08:48:27.821649 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znb9z" event={"ID":"b33b901e-2d7b-4f89-91b2-44615123ea7f","Type":"ContainerDied","Data":"021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8"} Oct 02 08:48:28 crc kubenswrapper[4960]: I1002 08:48:28.848328 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znb9z" event={"ID":"b33b901e-2d7b-4f89-91b2-44615123ea7f","Type":"ContainerStarted","Data":"e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75"} Oct 02 08:48:28 crc kubenswrapper[4960]: I1002 08:48:28.869864 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-znb9z" podStartSLOduration=3.457974883 podStartE2EDuration="6.869837302s" podCreationTimestamp="2025-10-02 08:48:22 +0000 UTC" firstStartedPulling="2025-10-02 08:48:24.794723353 +0000 UTC m=+5525.826669640" lastFinishedPulling="2025-10-02 08:48:28.206585772 +0000 UTC m=+5529.238532059" observedRunningTime="2025-10-02 08:48:28.867300313 +0000 UTC m=+5529.899246600" watchObservedRunningTime="2025-10-02 08:48:28.869837302 +0000 UTC m=+5529.901783589" Oct 02 08:48:33 crc kubenswrapper[4960]: I1002 08:48:33.300643 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:33 crc kubenswrapper[4960]: I1002 08:48:33.302960 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:33 crc kubenswrapper[4960]: I1002 08:48:33.357734 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:33 crc kubenswrapper[4960]: I1002 08:48:33.973108 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:34 crc kubenswrapper[4960]: I1002 08:48:34.021045 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-znb9z"] Oct 02 08:48:35 crc kubenswrapper[4960]: I1002 08:48:35.922549 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-znb9z" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="registry-server" containerID="cri-o://e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75" gracePeriod=2 Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.572051 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.646218 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-utilities\") pod \"b33b901e-2d7b-4f89-91b2-44615123ea7f\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.646499 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-catalog-content\") pod \"b33b901e-2d7b-4f89-91b2-44615123ea7f\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.646682 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfglr\" (UniqueName: \"kubernetes.io/projected/b33b901e-2d7b-4f89-91b2-44615123ea7f-kube-api-access-kfglr\") pod \"b33b901e-2d7b-4f89-91b2-44615123ea7f\" (UID: \"b33b901e-2d7b-4f89-91b2-44615123ea7f\") " Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.647818 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-utilities" (OuterVolumeSpecName: "utilities") pod "b33b901e-2d7b-4f89-91b2-44615123ea7f" (UID: "b33b901e-2d7b-4f89-91b2-44615123ea7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.657224 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b33b901e-2d7b-4f89-91b2-44615123ea7f-kube-api-access-kfglr" (OuterVolumeSpecName: "kube-api-access-kfglr") pod "b33b901e-2d7b-4f89-91b2-44615123ea7f" (UID: "b33b901e-2d7b-4f89-91b2-44615123ea7f"). InnerVolumeSpecName "kube-api-access-kfglr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.721809 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b33b901e-2d7b-4f89-91b2-44615123ea7f" (UID: "b33b901e-2d7b-4f89-91b2-44615123ea7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.751614 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.752064 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33b901e-2d7b-4f89-91b2-44615123ea7f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.752085 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfglr\" (UniqueName: \"kubernetes.io/projected/b33b901e-2d7b-4f89-91b2-44615123ea7f-kube-api-access-kfglr\") on node \"crc\" DevicePath \"\"" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.937009 4960 generic.go:334] "Generic (PLEG): container finished" podID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerID="e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75" exitCode=0 Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.937041 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znb9z" event={"ID":"b33b901e-2d7b-4f89-91b2-44615123ea7f","Type":"ContainerDied","Data":"e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75"} Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.937083 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znb9z" event={"ID":"b33b901e-2d7b-4f89-91b2-44615123ea7f","Type":"ContainerDied","Data":"e14a53d4e83a76065d0fc5b68171a0b4f2504050d9003f86815ce43f6de001b7"} Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.937101 4960 scope.go:117] "RemoveContainer" containerID="e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.937278 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znb9z" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.962740 4960 scope.go:117] "RemoveContainer" containerID="021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8" Oct 02 08:48:36 crc kubenswrapper[4960]: I1002 08:48:36.991822 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-znb9z"] Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.000327 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-znb9z"] Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.014884 4960 scope.go:117] "RemoveContainer" containerID="f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd" Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.050059 4960 scope.go:117] "RemoveContainer" containerID="e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75" Oct 02 08:48:37 crc kubenswrapper[4960]: E1002 08:48:37.051275 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75\": container with ID starting with e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75 not found: ID does not exist" containerID="e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75" Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.051322 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75"} err="failed to get container status \"e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75\": rpc error: code = NotFound desc = could not find container \"e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75\": container with ID starting with e3a3733732aece22a7bde56fdac375af7b7f9a0f517f0fff2be79b51c12f1c75 not found: ID does not exist" Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.051353 4960 scope.go:117] "RemoveContainer" containerID="021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8" Oct 02 08:48:37 crc kubenswrapper[4960]: E1002 08:48:37.051931 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8\": container with ID starting with 021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8 not found: ID does not exist" containerID="021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8" Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.052004 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8"} err="failed to get container status \"021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8\": rpc error: code = NotFound desc = could not find container \"021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8\": container with ID starting with 021319c11c667704edcf3547ef4cf09e3e78cce677df10dfc36ae31aefbd33d8 not found: ID does not exist" Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.052044 4960 scope.go:117] "RemoveContainer" containerID="f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd" Oct 02 08:48:37 crc kubenswrapper[4960]: E1002 08:48:37.052731 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd\": container with ID starting with f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd not found: ID does not exist" containerID="f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd" Oct 02 08:48:37 crc kubenswrapper[4960]: I1002 08:48:37.052767 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd"} err="failed to get container status \"f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd\": rpc error: code = NotFound desc = could not find container \"f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd\": container with ID starting with f18cd999932c77edceda1a106d6668a8ad8f849192fb102dacdd85e88a1740fd not found: ID does not exist" Oct 02 08:48:38 crc kubenswrapper[4960]: I1002 08:48:38.345643 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" path="/var/lib/kubelet/pods/b33b901e-2d7b-4f89-91b2-44615123ea7f/volumes" Oct 02 08:49:59 crc kubenswrapper[4960]: I1002 08:49:59.150067 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:49:59 crc kubenswrapper[4960]: I1002 08:49:59.150694 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:50:29 crc kubenswrapper[4960]: I1002 08:50:29.149633 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:50:29 crc kubenswrapper[4960]: I1002 08:50:29.150298 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.150589 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.151180 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.151242 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.152373 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.152445 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" gracePeriod=600 Oct 02 08:50:59 crc kubenswrapper[4960]: E1002 08:50:59.284699 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.407904 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" exitCode=0 Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.407954 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49"} Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.408046 4960 scope.go:117] "RemoveContainer" containerID="0c87b397fcffb54747129e24e768259dab6847d138d75073f9845299de9429a7" Oct 02 08:50:59 crc kubenswrapper[4960]: I1002 08:50:59.408735 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:50:59 crc kubenswrapper[4960]: E1002 08:50:59.409023 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:51:13 crc kubenswrapper[4960]: I1002 08:51:13.330513 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:51:13 crc kubenswrapper[4960]: E1002 08:51:13.331424 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:51:27 crc kubenswrapper[4960]: I1002 08:51:27.331515 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:51:27 crc kubenswrapper[4960]: E1002 08:51:27.332887 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:51:38 crc kubenswrapper[4960]: I1002 08:51:38.330358 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:51:38 crc kubenswrapper[4960]: E1002 08:51:38.331896 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:51:51 crc kubenswrapper[4960]: I1002 08:51:51.330204 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:51:51 crc kubenswrapper[4960]: E1002 08:51:51.331301 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:52:04 crc kubenswrapper[4960]: I1002 08:52:04.330454 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:52:04 crc kubenswrapper[4960]: E1002 08:52:04.331394 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:52:17 crc kubenswrapper[4960]: I1002 08:52:17.330525 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:52:17 crc kubenswrapper[4960]: E1002 08:52:17.331482 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:52:32 crc kubenswrapper[4960]: I1002 08:52:32.330568 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:52:32 crc kubenswrapper[4960]: E1002 08:52:32.331783 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:52:44 crc kubenswrapper[4960]: I1002 08:52:44.330491 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:52:44 crc kubenswrapper[4960]: E1002 08:52:44.331898 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:52:58 crc kubenswrapper[4960]: I1002 08:52:58.336213 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:52:58 crc kubenswrapper[4960]: E1002 08:52:58.337890 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:53:09 crc kubenswrapper[4960]: I1002 08:53:09.330649 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:53:09 crc kubenswrapper[4960]: E1002 08:53:09.332055 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:53:22 crc kubenswrapper[4960]: I1002 08:53:22.330791 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:53:22 crc kubenswrapper[4960]: E1002 08:53:22.332046 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:53:33 crc kubenswrapper[4960]: I1002 08:53:33.330655 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:53:33 crc kubenswrapper[4960]: E1002 08:53:33.331664 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:53:48 crc kubenswrapper[4960]: I1002 08:53:48.330562 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:53:48 crc kubenswrapper[4960]: E1002 08:53:48.331415 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:54:01 crc kubenswrapper[4960]: I1002 08:54:01.331791 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:54:01 crc kubenswrapper[4960]: E1002 08:54:01.332952 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:54:15 crc kubenswrapper[4960]: I1002 08:54:15.331032 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:54:15 crc kubenswrapper[4960]: E1002 08:54:15.332570 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:54:30 crc kubenswrapper[4960]: I1002 08:54:30.346254 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:54:30 crc kubenswrapper[4960]: E1002 08:54:30.349245 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:54:41 crc kubenswrapper[4960]: I1002 08:54:41.331327 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:54:41 crc kubenswrapper[4960]: E1002 08:54:41.332951 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:54:53 crc kubenswrapper[4960]: I1002 08:54:53.330566 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:54:53 crc kubenswrapper[4960]: E1002 08:54:53.332333 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:55:04 crc kubenswrapper[4960]: I1002 08:55:04.330333 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:55:04 crc kubenswrapper[4960]: E1002 08:55:04.331108 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:55:16 crc kubenswrapper[4960]: I1002 08:55:16.330180 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:55:16 crc kubenswrapper[4960]: E1002 08:55:16.331119 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:55:30 crc kubenswrapper[4960]: I1002 08:55:30.338505 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:55:30 crc kubenswrapper[4960]: E1002 08:55:30.339687 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:55:45 crc kubenswrapper[4960]: I1002 08:55:45.331538 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:55:45 crc kubenswrapper[4960]: E1002 08:55:45.333296 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:55:57 crc kubenswrapper[4960]: I1002 08:55:57.330233 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:55:57 crc kubenswrapper[4960]: E1002 08:55:57.331367 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 08:56:08 crc kubenswrapper[4960]: I1002 08:56:08.335638 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 08:56:08 crc kubenswrapper[4960]: I1002 08:56:08.837376 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"ab28f7b88f549884ae6b000b78dc914cf056e96b31108aa2eb74737cc774c097"} Oct 02 08:57:47 crc kubenswrapper[4960]: I1002 08:57:47.898171 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9g55n"] Oct 02 08:57:47 crc kubenswrapper[4960]: E1002 08:57:47.900024 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="extract-utilities" Oct 02 08:57:47 crc kubenswrapper[4960]: I1002 08:57:47.900053 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="extract-utilities" Oct 02 08:57:47 crc kubenswrapper[4960]: E1002 08:57:47.900109 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="extract-content" Oct 02 08:57:47 crc kubenswrapper[4960]: I1002 08:57:47.900123 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="extract-content" Oct 02 08:57:47 crc kubenswrapper[4960]: E1002 08:57:47.900137 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="registry-server" Oct 02 08:57:47 crc kubenswrapper[4960]: I1002 08:57:47.900149 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="registry-server" Oct 02 08:57:47 crc kubenswrapper[4960]: I1002 08:57:47.900550 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33b901e-2d7b-4f89-91b2-44615123ea7f" containerName="registry-server" Oct 02 08:57:47 crc kubenswrapper[4960]: I1002 08:57:47.903134 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:47 crc kubenswrapper[4960]: I1002 08:57:47.911822 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9g55n"] Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.062096 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-utilities\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.062171 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcqz8\" (UniqueName: \"kubernetes.io/projected/e6fbc015-8ee4-48ad-84f4-f8af553a2590-kube-api-access-mcqz8\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.062213 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-catalog-content\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.164266 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-utilities\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.164324 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcqz8\" (UniqueName: \"kubernetes.io/projected/e6fbc015-8ee4-48ad-84f4-f8af553a2590-kube-api-access-mcqz8\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.164364 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-catalog-content\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.164874 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-utilities\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.164960 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-catalog-content\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.193757 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcqz8\" (UniqueName: \"kubernetes.io/projected/e6fbc015-8ee4-48ad-84f4-f8af553a2590-kube-api-access-mcqz8\") pod \"certified-operators-9g55n\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.249600 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.838717 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9g55n"] Oct 02 08:57:48 crc kubenswrapper[4960]: I1002 08:57:48.986676 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9g55n" event={"ID":"e6fbc015-8ee4-48ad-84f4-f8af553a2590","Type":"ContainerStarted","Data":"87fd16510f2f845b6249ef704c475c4018295715c36b2e0a94f01a60ff4adb53"} Oct 02 08:57:50 crc kubenswrapper[4960]: I1002 08:57:50.004129 4960 generic.go:334] "Generic (PLEG): container finished" podID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerID="4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d" exitCode=0 Oct 02 08:57:50 crc kubenswrapper[4960]: I1002 08:57:50.004210 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9g55n" event={"ID":"e6fbc015-8ee4-48ad-84f4-f8af553a2590","Type":"ContainerDied","Data":"4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d"} Oct 02 08:57:50 crc kubenswrapper[4960]: I1002 08:57:50.007308 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:57:52 crc kubenswrapper[4960]: I1002 08:57:52.031871 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9g55n" event={"ID":"e6fbc015-8ee4-48ad-84f4-f8af553a2590","Type":"ContainerStarted","Data":"606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9"} Oct 02 08:57:53 crc kubenswrapper[4960]: I1002 08:57:53.047505 4960 generic.go:334] "Generic (PLEG): container finished" podID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerID="606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9" exitCode=0 Oct 02 08:57:53 crc kubenswrapper[4960]: I1002 08:57:53.047583 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9g55n" event={"ID":"e6fbc015-8ee4-48ad-84f4-f8af553a2590","Type":"ContainerDied","Data":"606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9"} Oct 02 08:57:55 crc kubenswrapper[4960]: I1002 08:57:55.078231 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9g55n" event={"ID":"e6fbc015-8ee4-48ad-84f4-f8af553a2590","Type":"ContainerStarted","Data":"82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80"} Oct 02 08:57:55 crc kubenswrapper[4960]: I1002 08:57:55.121412 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9g55n" podStartSLOduration=4.301490939 podStartE2EDuration="8.121387819s" podCreationTimestamp="2025-10-02 08:57:47 +0000 UTC" firstStartedPulling="2025-10-02 08:57:50.006899583 +0000 UTC m=+6091.038845880" lastFinishedPulling="2025-10-02 08:57:53.826796473 +0000 UTC m=+6094.858742760" observedRunningTime="2025-10-02 08:57:55.114443083 +0000 UTC m=+6096.146389440" watchObservedRunningTime="2025-10-02 08:57:55.121387819 +0000 UTC m=+6096.153334106" Oct 02 08:57:58 crc kubenswrapper[4960]: I1002 08:57:58.250185 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:58 crc kubenswrapper[4960]: I1002 08:57:58.250666 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:58 crc kubenswrapper[4960]: I1002 08:57:58.344435 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:59 crc kubenswrapper[4960]: I1002 08:57:59.210307 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:57:59 crc kubenswrapper[4960]: I1002 08:57:59.287499 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9g55n"] Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.149327 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9g55n" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="registry-server" containerID="cri-o://82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80" gracePeriod=2 Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.827046 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.847389 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-catalog-content\") pod \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.847480 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcqz8\" (UniqueName: \"kubernetes.io/projected/e6fbc015-8ee4-48ad-84f4-f8af553a2590-kube-api-access-mcqz8\") pod \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.848558 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-utilities\") pod \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\" (UID: \"e6fbc015-8ee4-48ad-84f4-f8af553a2590\") " Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.850012 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-utilities" (OuterVolumeSpecName: "utilities") pod "e6fbc015-8ee4-48ad-84f4-f8af553a2590" (UID: "e6fbc015-8ee4-48ad-84f4-f8af553a2590"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.851106 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.860030 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6fbc015-8ee4-48ad-84f4-f8af553a2590-kube-api-access-mcqz8" (OuterVolumeSpecName: "kube-api-access-mcqz8") pod "e6fbc015-8ee4-48ad-84f4-f8af553a2590" (UID: "e6fbc015-8ee4-48ad-84f4-f8af553a2590"). InnerVolumeSpecName "kube-api-access-mcqz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:58:01 crc kubenswrapper[4960]: I1002 08:58:01.952823 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcqz8\" (UniqueName: \"kubernetes.io/projected/e6fbc015-8ee4-48ad-84f4-f8af553a2590-kube-api-access-mcqz8\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.169665 4960 generic.go:334] "Generic (PLEG): container finished" podID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerID="82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80" exitCode=0 Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.169984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9g55n" event={"ID":"e6fbc015-8ee4-48ad-84f4-f8af553a2590","Type":"ContainerDied","Data":"82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80"} Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.170186 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9g55n" event={"ID":"e6fbc015-8ee4-48ad-84f4-f8af553a2590","Type":"ContainerDied","Data":"87fd16510f2f845b6249ef704c475c4018295715c36b2e0a94f01a60ff4adb53"} Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.170046 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9g55n" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.170222 4960 scope.go:117] "RemoveContainer" containerID="82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.195374 4960 scope.go:117] "RemoveContainer" containerID="606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.248710 4960 scope.go:117] "RemoveContainer" containerID="4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.279522 4960 scope.go:117] "RemoveContainer" containerID="82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80" Oct 02 08:58:02 crc kubenswrapper[4960]: E1002 08:58:02.280120 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80\": container with ID starting with 82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80 not found: ID does not exist" containerID="82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.280170 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80"} err="failed to get container status \"82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80\": rpc error: code = NotFound desc = could not find container \"82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80\": container with ID starting with 82fa12e138b6e136e72b19ed194e62b571b735a8481a0ff8ed5530085c10ab80 not found: ID does not exist" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.280208 4960 scope.go:117] "RemoveContainer" containerID="606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9" Oct 02 08:58:02 crc kubenswrapper[4960]: E1002 08:58:02.280464 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9\": container with ID starting with 606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9 not found: ID does not exist" containerID="606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.280497 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9"} err="failed to get container status \"606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9\": rpc error: code = NotFound desc = could not find container \"606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9\": container with ID starting with 606c40e26a968443b2030cc2beb78746e1094af7d673a3909e431404587d77c9 not found: ID does not exist" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.280552 4960 scope.go:117] "RemoveContainer" containerID="4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d" Oct 02 08:58:02 crc kubenswrapper[4960]: E1002 08:58:02.280758 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d\": container with ID starting with 4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d not found: ID does not exist" containerID="4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.280793 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d"} err="failed to get container status \"4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d\": rpc error: code = NotFound desc = could not find container \"4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d\": container with ID starting with 4a5a0f0257f44028d1e1ac62df1e64eaa60736364408c150a40171398eda635d not found: ID does not exist" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.409252 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6fbc015-8ee4-48ad-84f4-f8af553a2590" (UID: "e6fbc015-8ee4-48ad-84f4-f8af553a2590"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.477386 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6fbc015-8ee4-48ad-84f4-f8af553a2590-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.526332 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9g55n"] Oct 02 08:58:02 crc kubenswrapper[4960]: I1002 08:58:02.545788 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9g55n"] Oct 02 08:58:04 crc kubenswrapper[4960]: I1002 08:58:04.344573 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" path="/var/lib/kubelet/pods/e6fbc015-8ee4-48ad-84f4-f8af553a2590/volumes" Oct 02 08:58:29 crc kubenswrapper[4960]: I1002 08:58:29.150180 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:58:29 crc kubenswrapper[4960]: I1002 08:58:29.151151 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.376866 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lwkmg"] Oct 02 08:58:38 crc kubenswrapper[4960]: E1002 08:58:38.378139 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="extract-content" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.378159 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="extract-content" Oct 02 08:58:38 crc kubenswrapper[4960]: E1002 08:58:38.378198 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="registry-server" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.378210 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="registry-server" Oct 02 08:58:38 crc kubenswrapper[4960]: E1002 08:58:38.378252 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="extract-utilities" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.378262 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="extract-utilities" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.378556 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6fbc015-8ee4-48ad-84f4-f8af553a2590" containerName="registry-server" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.381744 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.402237 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lwkmg"] Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.482522 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6szq\" (UniqueName: \"kubernetes.io/projected/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-kube-api-access-p6szq\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.482644 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-utilities\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.482673 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-catalog-content\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.584477 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6szq\" (UniqueName: \"kubernetes.io/projected/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-kube-api-access-p6szq\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.584569 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-utilities\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.584594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-catalog-content\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.585338 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-catalog-content\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.585407 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-utilities\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.607411 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6szq\" (UniqueName: \"kubernetes.io/projected/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-kube-api-access-p6szq\") pod \"community-operators-lwkmg\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:38 crc kubenswrapper[4960]: I1002 08:58:38.706513 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:39 crc kubenswrapper[4960]: I1002 08:58:39.320948 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lwkmg"] Oct 02 08:58:39 crc kubenswrapper[4960]: I1002 08:58:39.568583 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerStarted","Data":"152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517"} Oct 02 08:58:39 crc kubenswrapper[4960]: I1002 08:58:39.569333 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerStarted","Data":"1e9a96f588f112fb916430fe4984480679ae4bee7f79c56016a9b0e7956d7d04"} Oct 02 08:58:40 crc kubenswrapper[4960]: I1002 08:58:40.584933 4960 generic.go:334] "Generic (PLEG): container finished" podID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerID="152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517" exitCode=0 Oct 02 08:58:40 crc kubenswrapper[4960]: I1002 08:58:40.585151 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerDied","Data":"152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517"} Oct 02 08:58:41 crc kubenswrapper[4960]: I1002 08:58:41.640747 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerStarted","Data":"c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025"} Oct 02 08:58:42 crc kubenswrapper[4960]: I1002 08:58:42.651532 4960 generic.go:334] "Generic (PLEG): container finished" podID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerID="c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025" exitCode=0 Oct 02 08:58:42 crc kubenswrapper[4960]: I1002 08:58:42.651640 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerDied","Data":"c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025"} Oct 02 08:58:43 crc kubenswrapper[4960]: I1002 08:58:43.665122 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerStarted","Data":"e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61"} Oct 02 08:58:43 crc kubenswrapper[4960]: I1002 08:58:43.707137 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lwkmg" podStartSLOduration=3.208000565 podStartE2EDuration="5.707100166s" podCreationTimestamp="2025-10-02 08:58:38 +0000 UTC" firstStartedPulling="2025-10-02 08:58:40.588691037 +0000 UTC m=+6141.620637354" lastFinishedPulling="2025-10-02 08:58:43.087790668 +0000 UTC m=+6144.119736955" observedRunningTime="2025-10-02 08:58:43.691959231 +0000 UTC m=+6144.723905518" watchObservedRunningTime="2025-10-02 08:58:43.707100166 +0000 UTC m=+6144.739046453" Oct 02 08:58:48 crc kubenswrapper[4960]: I1002 08:58:48.707593 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:48 crc kubenswrapper[4960]: I1002 08:58:48.708561 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:48 crc kubenswrapper[4960]: I1002 08:58:48.765942 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:48 crc kubenswrapper[4960]: I1002 08:58:48.828860 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:49 crc kubenswrapper[4960]: I1002 08:58:49.002960 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lwkmg"] Oct 02 08:58:50 crc kubenswrapper[4960]: I1002 08:58:50.749832 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lwkmg" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="registry-server" containerID="cri-o://e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61" gracePeriod=2 Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.344709 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.421653 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-utilities\") pod \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.421885 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6szq\" (UniqueName: \"kubernetes.io/projected/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-kube-api-access-p6szq\") pod \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.422138 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-catalog-content\") pod \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\" (UID: \"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56\") " Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.422430 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-utilities" (OuterVolumeSpecName: "utilities") pod "e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" (UID: "e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.422836 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.433263 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-kube-api-access-p6szq" (OuterVolumeSpecName: "kube-api-access-p6szq") pod "e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" (UID: "e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56"). InnerVolumeSpecName "kube-api-access-p6szq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.494638 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" (UID: "e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.525187 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.525537 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6szq\" (UniqueName: \"kubernetes.io/projected/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56-kube-api-access-p6szq\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.761621 4960 generic.go:334] "Generic (PLEG): container finished" podID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerID="e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61" exitCode=0 Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.761672 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerDied","Data":"e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61"} Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.761703 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lwkmg" event={"ID":"e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56","Type":"ContainerDied","Data":"1e9a96f588f112fb916430fe4984480679ae4bee7f79c56016a9b0e7956d7d04"} Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.761718 4960 scope.go:117] "RemoveContainer" containerID="e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.761822 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lwkmg" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.788599 4960 scope.go:117] "RemoveContainer" containerID="c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.822038 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lwkmg"] Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.826406 4960 scope.go:117] "RemoveContainer" containerID="152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.840615 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lwkmg"] Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.897415 4960 scope.go:117] "RemoveContainer" containerID="e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61" Oct 02 08:58:51 crc kubenswrapper[4960]: E1002 08:58:51.898575 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61\": container with ID starting with e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61 not found: ID does not exist" containerID="e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.898610 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61"} err="failed to get container status \"e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61\": rpc error: code = NotFound desc = could not find container \"e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61\": container with ID starting with e80c6ea3ba7b2b35b5a8a61df5f85543ebae4e805990fa2529a8009f8e0b5d61 not found: ID does not exist" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.898634 4960 scope.go:117] "RemoveContainer" containerID="c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025" Oct 02 08:58:51 crc kubenswrapper[4960]: E1002 08:58:51.898938 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025\": container with ID starting with c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025 not found: ID does not exist" containerID="c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.898985 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025"} err="failed to get container status \"c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025\": rpc error: code = NotFound desc = could not find container \"c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025\": container with ID starting with c37acad045c929c07eb4b7a55d97910a4b6682750be85166127228c39ba0c025 not found: ID does not exist" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.899024 4960 scope.go:117] "RemoveContainer" containerID="152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517" Oct 02 08:58:51 crc kubenswrapper[4960]: E1002 08:58:51.899238 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517\": container with ID starting with 152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517 not found: ID does not exist" containerID="152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517" Oct 02 08:58:51 crc kubenswrapper[4960]: I1002 08:58:51.899257 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517"} err="failed to get container status \"152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517\": rpc error: code = NotFound desc = could not find container \"152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517\": container with ID starting with 152115f5491fca76f117867967e5cb348f796f6adaaabbc4d9a2af822b815517 not found: ID does not exist" Oct 02 08:58:52 crc kubenswrapper[4960]: I1002 08:58:52.346437 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" path="/var/lib/kubelet/pods/e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56/volumes" Oct 02 08:58:59 crc kubenswrapper[4960]: I1002 08:58:59.150164 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:58:59 crc kubenswrapper[4960]: I1002 08:58:59.151438 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:59:29 crc kubenswrapper[4960]: I1002 08:59:29.150707 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:59:29 crc kubenswrapper[4960]: I1002 08:59:29.151628 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:59:29 crc kubenswrapper[4960]: I1002 08:59:29.151705 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 08:59:29 crc kubenswrapper[4960]: I1002 08:59:29.152762 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab28f7b88f549884ae6b000b78dc914cf056e96b31108aa2eb74737cc774c097"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:59:29 crc kubenswrapper[4960]: I1002 08:59:29.152830 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://ab28f7b88f549884ae6b000b78dc914cf056e96b31108aa2eb74737cc774c097" gracePeriod=600 Oct 02 08:59:30 crc kubenswrapper[4960]: I1002 08:59:30.183791 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="ab28f7b88f549884ae6b000b78dc914cf056e96b31108aa2eb74737cc774c097" exitCode=0 Oct 02 08:59:30 crc kubenswrapper[4960]: I1002 08:59:30.183870 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"ab28f7b88f549884ae6b000b78dc914cf056e96b31108aa2eb74737cc774c097"} Oct 02 08:59:30 crc kubenswrapper[4960]: I1002 08:59:30.185457 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4"} Oct 02 08:59:30 crc kubenswrapper[4960]: I1002 08:59:30.185554 4960 scope.go:117] "RemoveContainer" containerID="e19d4d5a1b1fede52d59f8344c88f5bb2675c55413f31783db6147d4c4d1cd49" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.208141 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846"] Oct 02 09:00:00 crc kubenswrapper[4960]: E1002 09:00:00.209435 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="extract-utilities" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.209459 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="extract-utilities" Oct 02 09:00:00 crc kubenswrapper[4960]: E1002 09:00:00.209487 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="extract-content" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.209494 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="extract-content" Oct 02 09:00:00 crc kubenswrapper[4960]: E1002 09:00:00.209525 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="registry-server" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.209531 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="registry-server" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.209716 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ee0b30-3dac-4d74-ad2d-4381a8d2ee56" containerName="registry-server" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.210535 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.218398 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846"] Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.224630 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.225152 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.310692 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-config-volume\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.311095 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-secret-volume\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.311130 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qs9c\" (UniqueName: \"kubernetes.io/projected/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-kube-api-access-2qs9c\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.413917 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-config-volume\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.414083 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-secret-volume\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.414106 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qs9c\" (UniqueName: \"kubernetes.io/projected/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-kube-api-access-2qs9c\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.415498 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-config-volume\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.433274 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-secret-volume\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.433929 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qs9c\" (UniqueName: \"kubernetes.io/projected/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-kube-api-access-2qs9c\") pod \"collect-profiles-29323260-zl846\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:00 crc kubenswrapper[4960]: I1002 09:00:00.550819 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:01 crc kubenswrapper[4960]: I1002 09:00:01.036602 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846"] Oct 02 09:00:01 crc kubenswrapper[4960]: W1002 09:00:01.044952 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8abfb90f_27ca_4fb0_8bed_2c4a76ab38a7.slice/crio-43ec360d9796be85c78a0cb245ded205455597973c3d511fa496dc6b1ca664ee WatchSource:0}: Error finding container 43ec360d9796be85c78a0cb245ded205455597973c3d511fa496dc6b1ca664ee: Status 404 returned error can't find the container with id 43ec360d9796be85c78a0cb245ded205455597973c3d511fa496dc6b1ca664ee Oct 02 09:00:01 crc kubenswrapper[4960]: I1002 09:00:01.503296 4960 generic.go:334] "Generic (PLEG): container finished" podID="8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" containerID="fa1148cccb5afb54b95d40792734f039630bf557bf19a7c2068b9be73fc92411" exitCode=0 Oct 02 09:00:01 crc kubenswrapper[4960]: I1002 09:00:01.503562 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" event={"ID":"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7","Type":"ContainerDied","Data":"fa1148cccb5afb54b95d40792734f039630bf557bf19a7c2068b9be73fc92411"} Oct 02 09:00:01 crc kubenswrapper[4960]: I1002 09:00:01.503598 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" event={"ID":"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7","Type":"ContainerStarted","Data":"43ec360d9796be85c78a0cb245ded205455597973c3d511fa496dc6b1ca664ee"} Oct 02 09:00:02 crc kubenswrapper[4960]: I1002 09:00:02.921206 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.080424 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-secret-volume\") pod \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.080627 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qs9c\" (UniqueName: \"kubernetes.io/projected/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-kube-api-access-2qs9c\") pod \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.081890 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-config-volume\") pod \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\" (UID: \"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7\") " Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.082688 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-config-volume" (OuterVolumeSpecName: "config-volume") pod "8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" (UID: "8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.083490 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.089780 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-kube-api-access-2qs9c" (OuterVolumeSpecName: "kube-api-access-2qs9c") pod "8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" (UID: "8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7"). InnerVolumeSpecName "kube-api-access-2qs9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.091077 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" (UID: "8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.185572 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.185627 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qs9c\" (UniqueName: \"kubernetes.io/projected/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7-kube-api-access-2qs9c\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.527059 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" event={"ID":"8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7","Type":"ContainerDied","Data":"43ec360d9796be85c78a0cb245ded205455597973c3d511fa496dc6b1ca664ee"} Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.527116 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43ec360d9796be85c78a0cb245ded205455597973c3d511fa496dc6b1ca664ee" Oct 02 09:00:03 crc kubenswrapper[4960]: I1002 09:00:03.527149 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846" Oct 02 09:00:04 crc kubenswrapper[4960]: I1002 09:00:04.037833 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh"] Oct 02 09:00:04 crc kubenswrapper[4960]: I1002 09:00:04.054426 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-j25kh"] Oct 02 09:00:04 crc kubenswrapper[4960]: I1002 09:00:04.345594 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e85da6fb-ac9e-49f3-813c-2f249b9e44ec" path="/var/lib/kubelet/pods/e85da6fb-ac9e-49f3-813c-2f249b9e44ec/volumes" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.651654 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r6x9x"] Oct 02 09:00:25 crc kubenswrapper[4960]: E1002 09:00:25.654015 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" containerName="collect-profiles" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.654038 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" containerName="collect-profiles" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.654491 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" containerName="collect-profiles" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.658284 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.679068 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6x9x"] Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.818163 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmg9f\" (UniqueName: \"kubernetes.io/projected/253fed54-83d3-4ad2-9b53-ed7872a716e6-kube-api-access-vmg9f\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.818337 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-catalog-content\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.818415 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-utilities\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.854578 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-524m4"] Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.856750 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.881303 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-524m4"] Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.923389 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmg9f\" (UniqueName: \"kubernetes.io/projected/253fed54-83d3-4ad2-9b53-ed7872a716e6-kube-api-access-vmg9f\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.923490 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-catalog-content\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.923527 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-utilities\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.924128 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-utilities\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.924405 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-catalog-content\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.968320 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmg9f\" (UniqueName: \"kubernetes.io/projected/253fed54-83d3-4ad2-9b53-ed7872a716e6-kube-api-access-vmg9f\") pod \"redhat-marketplace-r6x9x\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:25 crc kubenswrapper[4960]: I1002 09:00:25.992733 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.025322 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-catalog-content\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.025406 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlbf5\" (UniqueName: \"kubernetes.io/projected/87f56885-6657-4f25-8b74-cab334f91ecb-kube-api-access-rlbf5\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.025517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-utilities\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.127764 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-utilities\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.128739 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-catalog-content\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.128865 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlbf5\" (UniqueName: \"kubernetes.io/projected/87f56885-6657-4f25-8b74-cab334f91ecb-kube-api-access-rlbf5\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.129323 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-catalog-content\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.128429 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-utilities\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.150343 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlbf5\" (UniqueName: \"kubernetes.io/projected/87f56885-6657-4f25-8b74-cab334f91ecb-kube-api-access-rlbf5\") pod \"redhat-operators-524m4\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.176722 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.619334 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6x9x"] Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.774508 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6x9x" event={"ID":"253fed54-83d3-4ad2-9b53-ed7872a716e6","Type":"ContainerStarted","Data":"7aa5209b16b0e6c4dc2be68ee3f05175358a2c67e33d9b9656775a5766be8a08"} Oct 02 09:00:26 crc kubenswrapper[4960]: I1002 09:00:26.864367 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-524m4"] Oct 02 09:00:26 crc kubenswrapper[4960]: W1002 09:00:26.869749 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87f56885_6657_4f25_8b74_cab334f91ecb.slice/crio-de5028f63f201a067e87a0ccf468cc31e5ebdd9bac3eee879bf001135f71976d WatchSource:0}: Error finding container de5028f63f201a067e87a0ccf468cc31e5ebdd9bac3eee879bf001135f71976d: Status 404 returned error can't find the container with id de5028f63f201a067e87a0ccf468cc31e5ebdd9bac3eee879bf001135f71976d Oct 02 09:00:27 crc kubenswrapper[4960]: I1002 09:00:27.789164 4960 generic.go:334] "Generic (PLEG): container finished" podID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerID="e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20" exitCode=0 Oct 02 09:00:27 crc kubenswrapper[4960]: I1002 09:00:27.789301 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6x9x" event={"ID":"253fed54-83d3-4ad2-9b53-ed7872a716e6","Type":"ContainerDied","Data":"e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20"} Oct 02 09:00:27 crc kubenswrapper[4960]: I1002 09:00:27.795234 4960 generic.go:334] "Generic (PLEG): container finished" podID="87f56885-6657-4f25-8b74-cab334f91ecb" containerID="3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4" exitCode=0 Oct 02 09:00:27 crc kubenswrapper[4960]: I1002 09:00:27.795291 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-524m4" event={"ID":"87f56885-6657-4f25-8b74-cab334f91ecb","Type":"ContainerDied","Data":"3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4"} Oct 02 09:00:27 crc kubenswrapper[4960]: I1002 09:00:27.795329 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-524m4" event={"ID":"87f56885-6657-4f25-8b74-cab334f91ecb","Type":"ContainerStarted","Data":"de5028f63f201a067e87a0ccf468cc31e5ebdd9bac3eee879bf001135f71976d"} Oct 02 09:00:29 crc kubenswrapper[4960]: I1002 09:00:29.824124 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6x9x" event={"ID":"253fed54-83d3-4ad2-9b53-ed7872a716e6","Type":"ContainerStarted","Data":"35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa"} Oct 02 09:00:29 crc kubenswrapper[4960]: I1002 09:00:29.829699 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-524m4" event={"ID":"87f56885-6657-4f25-8b74-cab334f91ecb","Type":"ContainerStarted","Data":"98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d"} Oct 02 09:00:31 crc kubenswrapper[4960]: I1002 09:00:31.866853 4960 generic.go:334] "Generic (PLEG): container finished" podID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerID="35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa" exitCode=0 Oct 02 09:00:31 crc kubenswrapper[4960]: I1002 09:00:31.867004 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6x9x" event={"ID":"253fed54-83d3-4ad2-9b53-ed7872a716e6","Type":"ContainerDied","Data":"35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa"} Oct 02 09:00:32 crc kubenswrapper[4960]: I1002 09:00:32.881271 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6x9x" event={"ID":"253fed54-83d3-4ad2-9b53-ed7872a716e6","Type":"ContainerStarted","Data":"48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0"} Oct 02 09:00:32 crc kubenswrapper[4960]: I1002 09:00:32.884812 4960 generic.go:334] "Generic (PLEG): container finished" podID="87f56885-6657-4f25-8b74-cab334f91ecb" containerID="98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d" exitCode=0 Oct 02 09:00:32 crc kubenswrapper[4960]: I1002 09:00:32.884845 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-524m4" event={"ID":"87f56885-6657-4f25-8b74-cab334f91ecb","Type":"ContainerDied","Data":"98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d"} Oct 02 09:00:32 crc kubenswrapper[4960]: I1002 09:00:32.910964 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r6x9x" podStartSLOduration=3.250166872 podStartE2EDuration="7.910929175s" podCreationTimestamp="2025-10-02 09:00:25 +0000 UTC" firstStartedPulling="2025-10-02 09:00:27.791717121 +0000 UTC m=+6248.823663408" lastFinishedPulling="2025-10-02 09:00:32.452479424 +0000 UTC m=+6253.484425711" observedRunningTime="2025-10-02 09:00:32.909669201 +0000 UTC m=+6253.941615498" watchObservedRunningTime="2025-10-02 09:00:32.910929175 +0000 UTC m=+6253.942875472" Oct 02 09:00:33 crc kubenswrapper[4960]: I1002 09:00:33.902536 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-524m4" event={"ID":"87f56885-6657-4f25-8b74-cab334f91ecb","Type":"ContainerStarted","Data":"a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329"} Oct 02 09:00:33 crc kubenswrapper[4960]: I1002 09:00:33.939837 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-524m4" podStartSLOduration=3.283168545 podStartE2EDuration="8.939810244s" podCreationTimestamp="2025-10-02 09:00:25 +0000 UTC" firstStartedPulling="2025-10-02 09:00:27.797226408 +0000 UTC m=+6248.829172695" lastFinishedPulling="2025-10-02 09:00:33.453868107 +0000 UTC m=+6254.485814394" observedRunningTime="2025-10-02 09:00:33.923363704 +0000 UTC m=+6254.955310001" watchObservedRunningTime="2025-10-02 09:00:33.939810244 +0000 UTC m=+6254.971756531" Oct 02 09:00:35 crc kubenswrapper[4960]: I1002 09:00:35.993394 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:35 crc kubenswrapper[4960]: I1002 09:00:35.994043 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:36 crc kubenswrapper[4960]: I1002 09:00:36.053290 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:36 crc kubenswrapper[4960]: I1002 09:00:36.177577 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:36 crc kubenswrapper[4960]: I1002 09:00:36.177998 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:37 crc kubenswrapper[4960]: I1002 09:00:37.259889 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-524m4" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="registry-server" probeResult="failure" output=< Oct 02 09:00:37 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 09:00:37 crc kubenswrapper[4960]: > Oct 02 09:00:46 crc kubenswrapper[4960]: I1002 09:00:46.091519 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:46 crc kubenswrapper[4960]: I1002 09:00:46.143012 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6x9x"] Oct 02 09:00:46 crc kubenswrapper[4960]: I1002 09:00:46.241772 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:46 crc kubenswrapper[4960]: I1002 09:00:46.294387 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.058200 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r6x9x" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="registry-server" containerID="cri-o://48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0" gracePeriod=2 Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.554870 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.657818 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-catalog-content\") pod \"253fed54-83d3-4ad2-9b53-ed7872a716e6\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.657904 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmg9f\" (UniqueName: \"kubernetes.io/projected/253fed54-83d3-4ad2-9b53-ed7872a716e6-kube-api-access-vmg9f\") pod \"253fed54-83d3-4ad2-9b53-ed7872a716e6\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.657932 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-utilities\") pod \"253fed54-83d3-4ad2-9b53-ed7872a716e6\" (UID: \"253fed54-83d3-4ad2-9b53-ed7872a716e6\") " Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.658778 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-utilities" (OuterVolumeSpecName: "utilities") pod "253fed54-83d3-4ad2-9b53-ed7872a716e6" (UID: "253fed54-83d3-4ad2-9b53-ed7872a716e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.670338 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/253fed54-83d3-4ad2-9b53-ed7872a716e6-kube-api-access-vmg9f" (OuterVolumeSpecName: "kube-api-access-vmg9f") pod "253fed54-83d3-4ad2-9b53-ed7872a716e6" (UID: "253fed54-83d3-4ad2-9b53-ed7872a716e6"). InnerVolumeSpecName "kube-api-access-vmg9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.671227 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "253fed54-83d3-4ad2-9b53-ed7872a716e6" (UID: "253fed54-83d3-4ad2-9b53-ed7872a716e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.731106 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-524m4"] Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.760135 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.760175 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmg9f\" (UniqueName: \"kubernetes.io/projected/253fed54-83d3-4ad2-9b53-ed7872a716e6-kube-api-access-vmg9f\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:47 crc kubenswrapper[4960]: I1002 09:00:47.760186 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253fed54-83d3-4ad2-9b53-ed7872a716e6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.069804 4960 generic.go:334] "Generic (PLEG): container finished" podID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerID="48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0" exitCode=0 Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.069867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6x9x" event={"ID":"253fed54-83d3-4ad2-9b53-ed7872a716e6","Type":"ContainerDied","Data":"48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0"} Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.070211 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r6x9x" event={"ID":"253fed54-83d3-4ad2-9b53-ed7872a716e6","Type":"ContainerDied","Data":"7aa5209b16b0e6c4dc2be68ee3f05175358a2c67e33d9b9656775a5766be8a08"} Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.070238 4960 scope.go:117] "RemoveContainer" containerID="48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.069910 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r6x9x" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.070533 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-524m4" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="registry-server" containerID="cri-o://a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329" gracePeriod=2 Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.121900 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6x9x"] Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.123868 4960 scope.go:117] "RemoveContainer" containerID="35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.132183 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r6x9x"] Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.249714 4960 scope.go:117] "RemoveContainer" containerID="e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.296995 4960 scope.go:117] "RemoveContainer" containerID="48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0" Oct 02 09:00:48 crc kubenswrapper[4960]: E1002 09:00:48.297586 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0\": container with ID starting with 48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0 not found: ID does not exist" containerID="48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.297654 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0"} err="failed to get container status \"48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0\": rpc error: code = NotFound desc = could not find container \"48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0\": container with ID starting with 48ba3b1aaa3aae4102ecca7d672856051fbaca16dced4ae0cf6528649c2408d0 not found: ID does not exist" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.297701 4960 scope.go:117] "RemoveContainer" containerID="35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa" Oct 02 09:00:48 crc kubenswrapper[4960]: E1002 09:00:48.298184 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa\": container with ID starting with 35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa not found: ID does not exist" containerID="35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.298219 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa"} err="failed to get container status \"35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa\": rpc error: code = NotFound desc = could not find container \"35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa\": container with ID starting with 35ac20dee44763ad0e64c0d98940dd488c168eb5334c643cdc562e530b8681fa not found: ID does not exist" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.298248 4960 scope.go:117] "RemoveContainer" containerID="e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20" Oct 02 09:00:48 crc kubenswrapper[4960]: E1002 09:00:48.298487 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20\": container with ID starting with e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20 not found: ID does not exist" containerID="e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.298502 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20"} err="failed to get container status \"e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20\": rpc error: code = NotFound desc = could not find container \"e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20\": container with ID starting with e3c8ca9b0ae4f7f4503f5ee5b18f773803c8d8d8751996f9cebf893ab26a6e20 not found: ID does not exist" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.346142 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" path="/var/lib/kubelet/pods/253fed54-83d3-4ad2-9b53-ed7872a716e6/volumes" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.609395 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.780732 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-utilities\") pod \"87f56885-6657-4f25-8b74-cab334f91ecb\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.781315 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-catalog-content\") pod \"87f56885-6657-4f25-8b74-cab334f91ecb\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.781514 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlbf5\" (UniqueName: \"kubernetes.io/projected/87f56885-6657-4f25-8b74-cab334f91ecb-kube-api-access-rlbf5\") pod \"87f56885-6657-4f25-8b74-cab334f91ecb\" (UID: \"87f56885-6657-4f25-8b74-cab334f91ecb\") " Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.782201 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-utilities" (OuterVolumeSpecName: "utilities") pod "87f56885-6657-4f25-8b74-cab334f91ecb" (UID: "87f56885-6657-4f25-8b74-cab334f91ecb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.790002 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87f56885-6657-4f25-8b74-cab334f91ecb-kube-api-access-rlbf5" (OuterVolumeSpecName: "kube-api-access-rlbf5") pod "87f56885-6657-4f25-8b74-cab334f91ecb" (UID: "87f56885-6657-4f25-8b74-cab334f91ecb"). InnerVolumeSpecName "kube-api-access-rlbf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.882315 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87f56885-6657-4f25-8b74-cab334f91ecb" (UID: "87f56885-6657-4f25-8b74-cab334f91ecb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.884579 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.884610 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87f56885-6657-4f25-8b74-cab334f91ecb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:48 crc kubenswrapper[4960]: I1002 09:00:48.884629 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlbf5\" (UniqueName: \"kubernetes.io/projected/87f56885-6657-4f25-8b74-cab334f91ecb-kube-api-access-rlbf5\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.081514 4960 generic.go:334] "Generic (PLEG): container finished" podID="87f56885-6657-4f25-8b74-cab334f91ecb" containerID="a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329" exitCode=0 Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.081587 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-524m4" event={"ID":"87f56885-6657-4f25-8b74-cab334f91ecb","Type":"ContainerDied","Data":"a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329"} Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.081620 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-524m4" event={"ID":"87f56885-6657-4f25-8b74-cab334f91ecb","Type":"ContainerDied","Data":"de5028f63f201a067e87a0ccf468cc31e5ebdd9bac3eee879bf001135f71976d"} Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.081640 4960 scope.go:117] "RemoveContainer" containerID="a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.081763 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-524m4" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.118677 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-524m4"] Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.119121 4960 scope.go:117] "RemoveContainer" containerID="98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.126328 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-524m4"] Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.149331 4960 scope.go:117] "RemoveContainer" containerID="3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.173768 4960 scope.go:117] "RemoveContainer" containerID="a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329" Oct 02 09:00:49 crc kubenswrapper[4960]: E1002 09:00:49.174680 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329\": container with ID starting with a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329 not found: ID does not exist" containerID="a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.174721 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329"} err="failed to get container status \"a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329\": rpc error: code = NotFound desc = could not find container \"a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329\": container with ID starting with a2597591a59bf89e7640ddf29cb10f7dbe1468b99230643aff030ff016a97329 not found: ID does not exist" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.174768 4960 scope.go:117] "RemoveContainer" containerID="98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d" Oct 02 09:00:49 crc kubenswrapper[4960]: E1002 09:00:49.175291 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d\": container with ID starting with 98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d not found: ID does not exist" containerID="98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.175354 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d"} err="failed to get container status \"98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d\": rpc error: code = NotFound desc = could not find container \"98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d\": container with ID starting with 98b5c1a5c07449655e5d608f4ab1cc73839a16c30834d79fcfc31ca164e9d06d not found: ID does not exist" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.175398 4960 scope.go:117] "RemoveContainer" containerID="3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4" Oct 02 09:00:49 crc kubenswrapper[4960]: E1002 09:00:49.175751 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4\": container with ID starting with 3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4 not found: ID does not exist" containerID="3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4" Oct 02 09:00:49 crc kubenswrapper[4960]: I1002 09:00:49.175783 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4"} err="failed to get container status \"3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4\": rpc error: code = NotFound desc = could not find container \"3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4\": container with ID starting with 3d97f16aa1d7a2a57088c57c0e2be85ca91d0250172bd07329a62669708bdde4 not found: ID does not exist" Oct 02 09:00:50 crc kubenswrapper[4960]: I1002 09:00:50.344233 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" path="/var/lib/kubelet/pods/87f56885-6657-4f25-8b74-cab334f91ecb/volumes" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.163525 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323261-sp279"] Oct 02 09:01:00 crc kubenswrapper[4960]: E1002 09:01:00.164649 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="registry-server" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.164666 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="registry-server" Oct 02 09:01:00 crc kubenswrapper[4960]: E1002 09:01:00.164697 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="extract-content" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.164703 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="extract-content" Oct 02 09:01:00 crc kubenswrapper[4960]: E1002 09:01:00.164717 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="extract-utilities" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.164725 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="extract-utilities" Oct 02 09:01:00 crc kubenswrapper[4960]: E1002 09:01:00.164740 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="extract-content" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.164746 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="extract-content" Oct 02 09:01:00 crc kubenswrapper[4960]: E1002 09:01:00.164756 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="registry-server" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.164763 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="registry-server" Oct 02 09:01:00 crc kubenswrapper[4960]: E1002 09:01:00.164778 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="extract-utilities" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.164785 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="extract-utilities" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.165016 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="253fed54-83d3-4ad2-9b53-ed7872a716e6" containerName="registry-server" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.165045 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f56885-6657-4f25-8b74-cab334f91ecb" containerName="registry-server" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.165757 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.178557 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323261-sp279"] Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.249234 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-config-data\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.249372 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-combined-ca-bundle\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.249458 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwlj6\" (UniqueName: \"kubernetes.io/projected/aa245840-cc66-4671-a12c-00295fe7b5a4-kube-api-access-bwlj6\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.249482 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-fernet-keys\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.351527 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-combined-ca-bundle\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.351708 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwlj6\" (UniqueName: \"kubernetes.io/projected/aa245840-cc66-4671-a12c-00295fe7b5a4-kube-api-access-bwlj6\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.351740 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-fernet-keys\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.351809 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-config-data\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.358963 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-config-data\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.360331 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-fernet-keys\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.371092 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-combined-ca-bundle\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.375948 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwlj6\" (UniqueName: \"kubernetes.io/projected/aa245840-cc66-4671-a12c-00295fe7b5a4-kube-api-access-bwlj6\") pod \"keystone-cron-29323261-sp279\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.490024 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:00 crc kubenswrapper[4960]: I1002 09:01:00.959227 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323261-sp279"] Oct 02 09:01:01 crc kubenswrapper[4960]: I1002 09:01:01.205391 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-sp279" event={"ID":"aa245840-cc66-4671-a12c-00295fe7b5a4","Type":"ContainerStarted","Data":"b4a495b1a44803dad45e5590b4be5674f23d1890a7bb1e5b86308f6da6e15893"} Oct 02 09:01:02 crc kubenswrapper[4960]: I1002 09:01:02.216132 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-sp279" event={"ID":"aa245840-cc66-4671-a12c-00295fe7b5a4","Type":"ContainerStarted","Data":"6883e2eedaced2b19e756ea0cab03f3612f3d726bb40446cb73ff2aa755bc8dd"} Oct 02 09:01:03 crc kubenswrapper[4960]: I1002 09:01:03.840916 4960 scope.go:117] "RemoveContainer" containerID="3477d3b78361f5d88ce543d6318ad3c220a6ed682bbe7e1e9c09e022d00cefdf" Oct 02 09:01:04 crc kubenswrapper[4960]: I1002 09:01:04.239098 4960 generic.go:334] "Generic (PLEG): container finished" podID="aa245840-cc66-4671-a12c-00295fe7b5a4" containerID="6883e2eedaced2b19e756ea0cab03f3612f3d726bb40446cb73ff2aa755bc8dd" exitCode=0 Oct 02 09:01:04 crc kubenswrapper[4960]: I1002 09:01:04.239195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-sp279" event={"ID":"aa245840-cc66-4671-a12c-00295fe7b5a4","Type":"ContainerDied","Data":"6883e2eedaced2b19e756ea0cab03f3612f3d726bb40446cb73ff2aa755bc8dd"} Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.641201 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.799087 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-fernet-keys\") pod \"aa245840-cc66-4671-a12c-00295fe7b5a4\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.799138 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-combined-ca-bundle\") pod \"aa245840-cc66-4671-a12c-00295fe7b5a4\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.799166 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-config-data\") pod \"aa245840-cc66-4671-a12c-00295fe7b5a4\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.799265 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwlj6\" (UniqueName: \"kubernetes.io/projected/aa245840-cc66-4671-a12c-00295fe7b5a4-kube-api-access-bwlj6\") pod \"aa245840-cc66-4671-a12c-00295fe7b5a4\" (UID: \"aa245840-cc66-4671-a12c-00295fe7b5a4\") " Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.807114 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "aa245840-cc66-4671-a12c-00295fe7b5a4" (UID: "aa245840-cc66-4671-a12c-00295fe7b5a4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.807410 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa245840-cc66-4671-a12c-00295fe7b5a4-kube-api-access-bwlj6" (OuterVolumeSpecName: "kube-api-access-bwlj6") pod "aa245840-cc66-4671-a12c-00295fe7b5a4" (UID: "aa245840-cc66-4671-a12c-00295fe7b5a4"). InnerVolumeSpecName "kube-api-access-bwlj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.857676 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa245840-cc66-4671-a12c-00295fe7b5a4" (UID: "aa245840-cc66-4671-a12c-00295fe7b5a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.883107 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-config-data" (OuterVolumeSpecName: "config-data") pod "aa245840-cc66-4671-a12c-00295fe7b5a4" (UID: "aa245840-cc66-4671-a12c-00295fe7b5a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.902134 4960 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.902167 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.902177 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa245840-cc66-4671-a12c-00295fe7b5a4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4960]: I1002 09:01:05.902189 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwlj6\" (UniqueName: \"kubernetes.io/projected/aa245840-cc66-4671-a12c-00295fe7b5a4-kube-api-access-bwlj6\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:06 crc kubenswrapper[4960]: I1002 09:01:06.259292 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-sp279" Oct 02 09:01:06 crc kubenswrapper[4960]: I1002 09:01:06.259198 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-sp279" event={"ID":"aa245840-cc66-4671-a12c-00295fe7b5a4","Type":"ContainerDied","Data":"b4a495b1a44803dad45e5590b4be5674f23d1890a7bb1e5b86308f6da6e15893"} Oct 02 09:01:06 crc kubenswrapper[4960]: I1002 09:01:06.270190 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4a495b1a44803dad45e5590b4be5674f23d1890a7bb1e5b86308f6da6e15893" Oct 02 09:01:29 crc kubenswrapper[4960]: I1002 09:01:29.150576 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:01:29 crc kubenswrapper[4960]: I1002 09:01:29.151404 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:01:59 crc kubenswrapper[4960]: I1002 09:01:59.150595 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:01:59 crc kubenswrapper[4960]: I1002 09:01:59.151558 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:02:29 crc kubenswrapper[4960]: I1002 09:02:29.149827 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:02:29 crc kubenswrapper[4960]: I1002 09:02:29.150472 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:02:29 crc kubenswrapper[4960]: I1002 09:02:29.150527 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:02:29 crc kubenswrapper[4960]: I1002 09:02:29.151299 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:02:29 crc kubenswrapper[4960]: I1002 09:02:29.151364 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" gracePeriod=600 Oct 02 09:02:29 crc kubenswrapper[4960]: E1002 09:02:29.290967 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:02:30 crc kubenswrapper[4960]: I1002 09:02:30.136784 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" exitCode=0 Oct 02 09:02:30 crc kubenswrapper[4960]: I1002 09:02:30.136836 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4"} Oct 02 09:02:30 crc kubenswrapper[4960]: I1002 09:02:30.136876 4960 scope.go:117] "RemoveContainer" containerID="ab28f7b88f549884ae6b000b78dc914cf056e96b31108aa2eb74737cc774c097" Oct 02 09:02:30 crc kubenswrapper[4960]: I1002 09:02:30.137598 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:02:30 crc kubenswrapper[4960]: E1002 09:02:30.137929 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:02:45 crc kubenswrapper[4960]: I1002 09:02:45.330707 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:02:45 crc kubenswrapper[4960]: E1002 09:02:45.332295 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:02:58 crc kubenswrapper[4960]: I1002 09:02:58.332204 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:02:58 crc kubenswrapper[4960]: E1002 09:02:58.333706 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:03:12 crc kubenswrapper[4960]: I1002 09:03:12.343464 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:03:12 crc kubenswrapper[4960]: E1002 09:03:12.344707 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:03:26 crc kubenswrapper[4960]: I1002 09:03:26.330121 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:03:26 crc kubenswrapper[4960]: E1002 09:03:26.331092 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:03:38 crc kubenswrapper[4960]: I1002 09:03:38.330699 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:03:38 crc kubenswrapper[4960]: E1002 09:03:38.331564 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:03:52 crc kubenswrapper[4960]: I1002 09:03:52.330371 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:03:52 crc kubenswrapper[4960]: E1002 09:03:52.331467 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:04:04 crc kubenswrapper[4960]: I1002 09:04:04.330628 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:04:04 crc kubenswrapper[4960]: E1002 09:04:04.332422 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:04:16 crc kubenswrapper[4960]: I1002 09:04:16.331856 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:04:16 crc kubenswrapper[4960]: E1002 09:04:16.332832 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:04:28 crc kubenswrapper[4960]: I1002 09:04:28.334578 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:04:28 crc kubenswrapper[4960]: E1002 09:04:28.335991 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:04:41 crc kubenswrapper[4960]: I1002 09:04:41.330073 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:04:41 crc kubenswrapper[4960]: E1002 09:04:41.330897 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:04:56 crc kubenswrapper[4960]: I1002 09:04:56.329730 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:04:56 crc kubenswrapper[4960]: E1002 09:04:56.330600 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:05:10 crc kubenswrapper[4960]: I1002 09:05:10.338805 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:05:10 crc kubenswrapper[4960]: E1002 09:05:10.339709 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:05:23 crc kubenswrapper[4960]: I1002 09:05:23.331687 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:05:23 crc kubenswrapper[4960]: E1002 09:05:23.333070 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:05:37 crc kubenswrapper[4960]: I1002 09:05:37.329793 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:05:37 crc kubenswrapper[4960]: E1002 09:05:37.330769 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:05:48 crc kubenswrapper[4960]: I1002 09:05:48.329914 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:05:48 crc kubenswrapper[4960]: E1002 09:05:48.330848 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:06:03 crc kubenswrapper[4960]: I1002 09:06:03.332286 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:06:03 crc kubenswrapper[4960]: E1002 09:06:03.333704 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:06:14 crc kubenswrapper[4960]: I1002 09:06:14.331435 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:06:14 crc kubenswrapper[4960]: E1002 09:06:14.332956 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:06:25 crc kubenswrapper[4960]: I1002 09:06:25.331135 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:06:25 crc kubenswrapper[4960]: E1002 09:06:25.332353 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:06:38 crc kubenswrapper[4960]: I1002 09:06:38.330882 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:06:38 crc kubenswrapper[4960]: E1002 09:06:38.332249 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:06:49 crc kubenswrapper[4960]: I1002 09:06:49.330433 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:06:49 crc kubenswrapper[4960]: E1002 09:06:49.331567 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:07:03 crc kubenswrapper[4960]: I1002 09:07:03.330542 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:07:03 crc kubenswrapper[4960]: E1002 09:07:03.331435 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:07:15 crc kubenswrapper[4960]: I1002 09:07:15.330219 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:07:15 crc kubenswrapper[4960]: E1002 09:07:15.331131 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:07:26 crc kubenswrapper[4960]: I1002 09:07:26.329908 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:07:26 crc kubenswrapper[4960]: E1002 09:07:26.330702 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:07:38 crc kubenswrapper[4960]: I1002 09:07:38.330536 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:07:39 crc kubenswrapper[4960]: I1002 09:07:39.401123 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"24fa21f7e874bc070325e0ed12f94c2337c713653061abc6301926e6dd32a6f6"} Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.300702 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bnjm7"] Oct 02 09:08:18 crc kubenswrapper[4960]: E1002 09:08:18.301876 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa245840-cc66-4671-a12c-00295fe7b5a4" containerName="keystone-cron" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.301888 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa245840-cc66-4671-a12c-00295fe7b5a4" containerName="keystone-cron" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.302097 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa245840-cc66-4671-a12c-00295fe7b5a4" containerName="keystone-cron" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.303473 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.312250 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bnjm7"] Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.434085 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-utilities\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.434232 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-catalog-content\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.434307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wqk6\" (UniqueName: \"kubernetes.io/projected/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-kube-api-access-9wqk6\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.537350 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-utilities\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.537479 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-catalog-content\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.537531 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wqk6\" (UniqueName: \"kubernetes.io/projected/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-kube-api-access-9wqk6\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.537952 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-utilities\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.538107 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-catalog-content\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.565805 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wqk6\" (UniqueName: \"kubernetes.io/projected/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-kube-api-access-9wqk6\") pod \"certified-operators-bnjm7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:18 crc kubenswrapper[4960]: I1002 09:08:18.644213 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:19 crc kubenswrapper[4960]: I1002 09:08:19.139718 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bnjm7"] Oct 02 09:08:19 crc kubenswrapper[4960]: I1002 09:08:19.789216 4960 generic.go:334] "Generic (PLEG): container finished" podID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerID="403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c" exitCode=0 Oct 02 09:08:19 crc kubenswrapper[4960]: I1002 09:08:19.789274 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnjm7" event={"ID":"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7","Type":"ContainerDied","Data":"403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c"} Oct 02 09:08:19 crc kubenswrapper[4960]: I1002 09:08:19.789324 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnjm7" event={"ID":"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7","Type":"ContainerStarted","Data":"e4663f97c8f480ef5392a12c92b2bab5963f536955b3ed5984c65729ee115960"} Oct 02 09:08:19 crc kubenswrapper[4960]: I1002 09:08:19.792172 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:08:21 crc kubenswrapper[4960]: I1002 09:08:21.813694 4960 generic.go:334] "Generic (PLEG): container finished" podID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerID="ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054" exitCode=0 Oct 02 09:08:21 crc kubenswrapper[4960]: I1002 09:08:21.814003 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnjm7" event={"ID":"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7","Type":"ContainerDied","Data":"ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054"} Oct 02 09:08:22 crc kubenswrapper[4960]: I1002 09:08:22.846545 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnjm7" event={"ID":"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7","Type":"ContainerStarted","Data":"7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760"} Oct 02 09:08:22 crc kubenswrapper[4960]: I1002 09:08:22.869247 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bnjm7" podStartSLOduration=2.320256693 podStartE2EDuration="4.869219859s" podCreationTimestamp="2025-10-02 09:08:18 +0000 UTC" firstStartedPulling="2025-10-02 09:08:19.79192814 +0000 UTC m=+6720.823874427" lastFinishedPulling="2025-10-02 09:08:22.340891306 +0000 UTC m=+6723.372837593" observedRunningTime="2025-10-02 09:08:22.865213641 +0000 UTC m=+6723.897159938" watchObservedRunningTime="2025-10-02 09:08:22.869219859 +0000 UTC m=+6723.901166176" Oct 02 09:08:28 crc kubenswrapper[4960]: I1002 09:08:28.645414 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:28 crc kubenswrapper[4960]: I1002 09:08:28.646017 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:28 crc kubenswrapper[4960]: I1002 09:08:28.693834 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:28 crc kubenswrapper[4960]: I1002 09:08:28.951327 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:29 crc kubenswrapper[4960]: I1002 09:08:29.000128 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bnjm7"] Oct 02 09:08:30 crc kubenswrapper[4960]: I1002 09:08:30.918952 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bnjm7" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="registry-server" containerID="cri-o://7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760" gracePeriod=2 Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.642086 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.662289 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-utilities\") pod \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.662340 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wqk6\" (UniqueName: \"kubernetes.io/projected/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-kube-api-access-9wqk6\") pod \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.662494 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-catalog-content\") pod \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\" (UID: \"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7\") " Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.663388 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-utilities" (OuterVolumeSpecName: "utilities") pod "e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" (UID: "e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.670307 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-kube-api-access-9wqk6" (OuterVolumeSpecName: "kube-api-access-9wqk6") pod "e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" (UID: "e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7"). InnerVolumeSpecName "kube-api-access-9wqk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.711835 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" (UID: "e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.765794 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.765838 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wqk6\" (UniqueName: \"kubernetes.io/projected/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-kube-api-access-9wqk6\") on node \"crc\" DevicePath \"\"" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.765853 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.930154 4960 generic.go:334] "Generic (PLEG): container finished" podID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerID="7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760" exitCode=0 Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.930227 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnjm7" event={"ID":"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7","Type":"ContainerDied","Data":"7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760"} Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.930303 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bnjm7" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.930345 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bnjm7" event={"ID":"e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7","Type":"ContainerDied","Data":"e4663f97c8f480ef5392a12c92b2bab5963f536955b3ed5984c65729ee115960"} Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.930395 4960 scope.go:117] "RemoveContainer" containerID="7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.952967 4960 scope.go:117] "RemoveContainer" containerID="ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054" Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.975900 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bnjm7"] Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.985076 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bnjm7"] Oct 02 09:08:31 crc kubenswrapper[4960]: I1002 09:08:31.996860 4960 scope.go:117] "RemoveContainer" containerID="403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c" Oct 02 09:08:32 crc kubenswrapper[4960]: I1002 09:08:32.039190 4960 scope.go:117] "RemoveContainer" containerID="7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760" Oct 02 09:08:32 crc kubenswrapper[4960]: E1002 09:08:32.039730 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760\": container with ID starting with 7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760 not found: ID does not exist" containerID="7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760" Oct 02 09:08:32 crc kubenswrapper[4960]: I1002 09:08:32.039781 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760"} err="failed to get container status \"7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760\": rpc error: code = NotFound desc = could not find container \"7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760\": container with ID starting with 7e51209dc919a224e42891747b880272d43caa9190bcb709aafbf43298c58760 not found: ID does not exist" Oct 02 09:08:32 crc kubenswrapper[4960]: I1002 09:08:32.039814 4960 scope.go:117] "RemoveContainer" containerID="ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054" Oct 02 09:08:32 crc kubenswrapper[4960]: E1002 09:08:32.040389 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054\": container with ID starting with ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054 not found: ID does not exist" containerID="ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054" Oct 02 09:08:32 crc kubenswrapper[4960]: I1002 09:08:32.040448 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054"} err="failed to get container status \"ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054\": rpc error: code = NotFound desc = could not find container \"ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054\": container with ID starting with ea276dce6066df4841fe03a6c72ff5f77d029abe0badbd7a0b276be87c711054 not found: ID does not exist" Oct 02 09:08:32 crc kubenswrapper[4960]: I1002 09:08:32.040484 4960 scope.go:117] "RemoveContainer" containerID="403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c" Oct 02 09:08:32 crc kubenswrapper[4960]: E1002 09:08:32.040850 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c\": container with ID starting with 403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c not found: ID does not exist" containerID="403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c" Oct 02 09:08:32 crc kubenswrapper[4960]: I1002 09:08:32.040906 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c"} err="failed to get container status \"403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c\": rpc error: code = NotFound desc = could not find container \"403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c\": container with ID starting with 403e02d493d4674176fe19fa6853e6a105b33416de95a51567cb230fe51fe16c not found: ID does not exist" Oct 02 09:08:32 crc kubenswrapper[4960]: I1002 09:08:32.352712 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" path="/var/lib/kubelet/pods/e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7/volumes" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.090357 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v5qkv"] Oct 02 09:09:46 crc kubenswrapper[4960]: E1002 09:09:46.092631 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="extract-utilities" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.092753 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="extract-utilities" Oct 02 09:09:46 crc kubenswrapper[4960]: E1002 09:09:46.092811 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="registry-server" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.092878 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="registry-server" Oct 02 09:09:46 crc kubenswrapper[4960]: E1002 09:09:46.092914 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="extract-content" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.092920 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="extract-content" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.093337 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d9a6ce-4dde-4307-8f17-1ccdf0e0d7f7" containerName="registry-server" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.097106 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.112850 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5qkv"] Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.185242 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-catalog-content\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.185328 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chmtg\" (UniqueName: \"kubernetes.io/projected/8b44e163-67b5-442e-a5cb-8078575984ae-kube-api-access-chmtg\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.185437 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-utilities\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.287663 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-utilities\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.287797 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-catalog-content\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.287838 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chmtg\" (UniqueName: \"kubernetes.io/projected/8b44e163-67b5-442e-a5cb-8078575984ae-kube-api-access-chmtg\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.288750 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-utilities\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.288898 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-catalog-content\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.332340 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chmtg\" (UniqueName: \"kubernetes.io/projected/8b44e163-67b5-442e-a5cb-8078575984ae-kube-api-access-chmtg\") pod \"community-operators-v5qkv\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:46 crc kubenswrapper[4960]: I1002 09:09:46.447774 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:47 crc kubenswrapper[4960]: I1002 09:09:47.109316 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5qkv"] Oct 02 09:09:47 crc kubenswrapper[4960]: I1002 09:09:47.761733 4960 generic.go:334] "Generic (PLEG): container finished" podID="8b44e163-67b5-442e-a5cb-8078575984ae" containerID="4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7" exitCode=0 Oct 02 09:09:47 crc kubenswrapper[4960]: I1002 09:09:47.761805 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5qkv" event={"ID":"8b44e163-67b5-442e-a5cb-8078575984ae","Type":"ContainerDied","Data":"4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7"} Oct 02 09:09:47 crc kubenswrapper[4960]: I1002 09:09:47.762232 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5qkv" event={"ID":"8b44e163-67b5-442e-a5cb-8078575984ae","Type":"ContainerStarted","Data":"b6801d1912443610b9d1dd1cdb6e2d6ae062496f4e6c42ca1acddf361236c38b"} Oct 02 09:09:49 crc kubenswrapper[4960]: I1002 09:09:49.793525 4960 generic.go:334] "Generic (PLEG): container finished" podID="8b44e163-67b5-442e-a5cb-8078575984ae" containerID="7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a" exitCode=0 Oct 02 09:09:49 crc kubenswrapper[4960]: I1002 09:09:49.793629 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5qkv" event={"ID":"8b44e163-67b5-442e-a5cb-8078575984ae","Type":"ContainerDied","Data":"7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a"} Oct 02 09:09:51 crc kubenswrapper[4960]: I1002 09:09:51.822114 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5qkv" event={"ID":"8b44e163-67b5-442e-a5cb-8078575984ae","Type":"ContainerStarted","Data":"5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf"} Oct 02 09:09:51 crc kubenswrapper[4960]: I1002 09:09:51.847854 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v5qkv" podStartSLOduration=4.003285814 podStartE2EDuration="6.847825097s" podCreationTimestamp="2025-10-02 09:09:45 +0000 UTC" firstStartedPulling="2025-10-02 09:09:47.767290886 +0000 UTC m=+6808.799237173" lastFinishedPulling="2025-10-02 09:09:50.611830169 +0000 UTC m=+6811.643776456" observedRunningTime="2025-10-02 09:09:51.840878271 +0000 UTC m=+6812.872824568" watchObservedRunningTime="2025-10-02 09:09:51.847825097 +0000 UTC m=+6812.879771384" Oct 02 09:09:56 crc kubenswrapper[4960]: I1002 09:09:56.448875 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:56 crc kubenswrapper[4960]: I1002 09:09:56.450020 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:56 crc kubenswrapper[4960]: I1002 09:09:56.505191 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:56 crc kubenswrapper[4960]: I1002 09:09:56.931161 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:56 crc kubenswrapper[4960]: I1002 09:09:56.999181 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5qkv"] Oct 02 09:09:58 crc kubenswrapper[4960]: I1002 09:09:58.901856 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v5qkv" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="registry-server" containerID="cri-o://5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf" gracePeriod=2 Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.152755 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.152918 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.461402 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.492190 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-catalog-content\") pod \"8b44e163-67b5-442e-a5cb-8078575984ae\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.492348 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chmtg\" (UniqueName: \"kubernetes.io/projected/8b44e163-67b5-442e-a5cb-8078575984ae-kube-api-access-chmtg\") pod \"8b44e163-67b5-442e-a5cb-8078575984ae\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.492456 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-utilities\") pod \"8b44e163-67b5-442e-a5cb-8078575984ae\" (UID: \"8b44e163-67b5-442e-a5cb-8078575984ae\") " Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.494266 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-utilities" (OuterVolumeSpecName: "utilities") pod "8b44e163-67b5-442e-a5cb-8078575984ae" (UID: "8b44e163-67b5-442e-a5cb-8078575984ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.505012 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b44e163-67b5-442e-a5cb-8078575984ae-kube-api-access-chmtg" (OuterVolumeSpecName: "kube-api-access-chmtg") pod "8b44e163-67b5-442e-a5cb-8078575984ae" (UID: "8b44e163-67b5-442e-a5cb-8078575984ae"). InnerVolumeSpecName "kube-api-access-chmtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.549506 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b44e163-67b5-442e-a5cb-8078575984ae" (UID: "8b44e163-67b5-442e-a5cb-8078575984ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.596131 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.596178 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chmtg\" (UniqueName: \"kubernetes.io/projected/8b44e163-67b5-442e-a5cb-8078575984ae-kube-api-access-chmtg\") on node \"crc\" DevicePath \"\"" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.596189 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b44e163-67b5-442e-a5cb-8078575984ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.921904 4960 generic.go:334] "Generic (PLEG): container finished" podID="8b44e163-67b5-442e-a5cb-8078575984ae" containerID="5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf" exitCode=0 Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.921995 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5qkv" event={"ID":"8b44e163-67b5-442e-a5cb-8078575984ae","Type":"ContainerDied","Data":"5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf"} Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.922042 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5qkv" event={"ID":"8b44e163-67b5-442e-a5cb-8078575984ae","Type":"ContainerDied","Data":"b6801d1912443610b9d1dd1cdb6e2d6ae062496f4e6c42ca1acddf361236c38b"} Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.922070 4960 scope.go:117] "RemoveContainer" containerID="5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.923818 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5qkv" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.954162 4960 scope.go:117] "RemoveContainer" containerID="7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a" Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.970996 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5qkv"] Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.982522 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v5qkv"] Oct 02 09:09:59 crc kubenswrapper[4960]: I1002 09:09:59.991642 4960 scope.go:117] "RemoveContainer" containerID="4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7" Oct 02 09:10:00 crc kubenswrapper[4960]: I1002 09:10:00.033479 4960 scope.go:117] "RemoveContainer" containerID="5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf" Oct 02 09:10:00 crc kubenswrapper[4960]: E1002 09:10:00.034689 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf\": container with ID starting with 5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf not found: ID does not exist" containerID="5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf" Oct 02 09:10:00 crc kubenswrapper[4960]: I1002 09:10:00.034752 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf"} err="failed to get container status \"5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf\": rpc error: code = NotFound desc = could not find container \"5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf\": container with ID starting with 5f8e75206db204321b1640442f39bbc32c34cb2215e157e483bd355b345ed0cf not found: ID does not exist" Oct 02 09:10:00 crc kubenswrapper[4960]: I1002 09:10:00.034789 4960 scope.go:117] "RemoveContainer" containerID="7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a" Oct 02 09:10:00 crc kubenswrapper[4960]: E1002 09:10:00.035419 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a\": container with ID starting with 7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a not found: ID does not exist" containerID="7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a" Oct 02 09:10:00 crc kubenswrapper[4960]: I1002 09:10:00.035478 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a"} err="failed to get container status \"7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a\": rpc error: code = NotFound desc = could not find container \"7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a\": container with ID starting with 7cb0f18f699b7b687ec9a3748c4db038d604dc082e20dd3778a555b6a124d73a not found: ID does not exist" Oct 02 09:10:00 crc kubenswrapper[4960]: I1002 09:10:00.035526 4960 scope.go:117] "RemoveContainer" containerID="4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7" Oct 02 09:10:00 crc kubenswrapper[4960]: E1002 09:10:00.036105 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7\": container with ID starting with 4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7 not found: ID does not exist" containerID="4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7" Oct 02 09:10:00 crc kubenswrapper[4960]: I1002 09:10:00.036155 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7"} err="failed to get container status \"4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7\": rpc error: code = NotFound desc = could not find container \"4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7\": container with ID starting with 4d759e01e871f3d6b32e11cc15a86e1ee3a8a4bccfda7692a103fe69bca743d7 not found: ID does not exist" Oct 02 09:10:00 crc kubenswrapper[4960]: I1002 09:10:00.344815 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" path="/var/lib/kubelet/pods/8b44e163-67b5-442e-a5cb-8078575984ae/volumes" Oct 02 09:10:29 crc kubenswrapper[4960]: I1002 09:10:29.150078 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:10:29 crc kubenswrapper[4960]: I1002 09:10:29.151375 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.365024 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xbcjd"] Oct 02 09:10:51 crc kubenswrapper[4960]: E1002 09:10:51.366506 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="extract-content" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.366526 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="extract-content" Oct 02 09:10:51 crc kubenswrapper[4960]: E1002 09:10:51.366542 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="extract-utilities" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.366553 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="extract-utilities" Oct 02 09:10:51 crc kubenswrapper[4960]: E1002 09:10:51.366567 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="registry-server" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.366574 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="registry-server" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.366816 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b44e163-67b5-442e-a5cb-8078575984ae" containerName="registry-server" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.368649 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.375013 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xbcjd"] Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.537068 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-utilities\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.537150 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4zmg\" (UniqueName: \"kubernetes.io/projected/b805a38d-5112-4a9d-b30b-59635596aacc-kube-api-access-q4zmg\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.537182 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-catalog-content\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.639791 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-utilities\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.639869 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4zmg\" (UniqueName: \"kubernetes.io/projected/b805a38d-5112-4a9d-b30b-59635596aacc-kube-api-access-q4zmg\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.639892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-catalog-content\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.640355 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-utilities\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.640477 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-catalog-content\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.660772 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4zmg\" (UniqueName: \"kubernetes.io/projected/b805a38d-5112-4a9d-b30b-59635596aacc-kube-api-access-q4zmg\") pod \"redhat-operators-xbcjd\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:51 crc kubenswrapper[4960]: I1002 09:10:51.687539 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:10:52 crc kubenswrapper[4960]: I1002 09:10:52.182837 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xbcjd"] Oct 02 09:10:52 crc kubenswrapper[4960]: I1002 09:10:52.531094 4960 generic.go:334] "Generic (PLEG): container finished" podID="b805a38d-5112-4a9d-b30b-59635596aacc" containerID="dec8e44a4ac5d90fe731d809995a0133ffe0690a5156169cd888566ce5760251" exitCode=0 Oct 02 09:10:52 crc kubenswrapper[4960]: I1002 09:10:52.531188 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbcjd" event={"ID":"b805a38d-5112-4a9d-b30b-59635596aacc","Type":"ContainerDied","Data":"dec8e44a4ac5d90fe731d809995a0133ffe0690a5156169cd888566ce5760251"} Oct 02 09:10:52 crc kubenswrapper[4960]: I1002 09:10:52.531580 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbcjd" event={"ID":"b805a38d-5112-4a9d-b30b-59635596aacc","Type":"ContainerStarted","Data":"9e0c7150188366a87358898a521a4c9bf893f0557bcb6ed1e18fdacfd895b001"} Oct 02 09:10:54 crc kubenswrapper[4960]: I1002 09:10:54.551594 4960 generic.go:334] "Generic (PLEG): container finished" podID="b805a38d-5112-4a9d-b30b-59635596aacc" containerID="8c14243de13686db01f255221b09b20370e7c90a59a0977a7b7b506ef6887f45" exitCode=0 Oct 02 09:10:54 crc kubenswrapper[4960]: I1002 09:10:54.551682 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbcjd" event={"ID":"b805a38d-5112-4a9d-b30b-59635596aacc","Type":"ContainerDied","Data":"8c14243de13686db01f255221b09b20370e7c90a59a0977a7b7b506ef6887f45"} Oct 02 09:10:55 crc kubenswrapper[4960]: I1002 09:10:55.566667 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbcjd" event={"ID":"b805a38d-5112-4a9d-b30b-59635596aacc","Type":"ContainerStarted","Data":"d96cc644bd52cdfa7b575afaf735b201dc58be52627009f86589430b25422bf6"} Oct 02 09:10:55 crc kubenswrapper[4960]: I1002 09:10:55.590146 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xbcjd" podStartSLOduration=2.129803946 podStartE2EDuration="4.590116558s" podCreationTimestamp="2025-10-02 09:10:51 +0000 UTC" firstStartedPulling="2025-10-02 09:10:52.533567545 +0000 UTC m=+6873.565513832" lastFinishedPulling="2025-10-02 09:10:54.993880157 +0000 UTC m=+6876.025826444" observedRunningTime="2025-10-02 09:10:55.584363574 +0000 UTC m=+6876.616309861" watchObservedRunningTime="2025-10-02 09:10:55.590116558 +0000 UTC m=+6876.622062845" Oct 02 09:10:59 crc kubenswrapper[4960]: I1002 09:10:59.150775 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:10:59 crc kubenswrapper[4960]: I1002 09:10:59.151392 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:10:59 crc kubenswrapper[4960]: I1002 09:10:59.151477 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:10:59 crc kubenswrapper[4960]: I1002 09:10:59.153197 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24fa21f7e874bc070325e0ed12f94c2337c713653061abc6301926e6dd32a6f6"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:10:59 crc kubenswrapper[4960]: I1002 09:10:59.153385 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://24fa21f7e874bc070325e0ed12f94c2337c713653061abc6301926e6dd32a6f6" gracePeriod=600 Oct 02 09:11:00 crc kubenswrapper[4960]: I1002 09:11:00.619264 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="24fa21f7e874bc070325e0ed12f94c2337c713653061abc6301926e6dd32a6f6" exitCode=0 Oct 02 09:11:00 crc kubenswrapper[4960]: I1002 09:11:00.619336 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"24fa21f7e874bc070325e0ed12f94c2337c713653061abc6301926e6dd32a6f6"} Oct 02 09:11:00 crc kubenswrapper[4960]: I1002 09:11:00.619832 4960 scope.go:117] "RemoveContainer" containerID="f88023ae05c30f53571215bf69c113d975fcbc076700916eb7d6698d7eae28a4" Oct 02 09:11:00 crc kubenswrapper[4960]: I1002 09:11:00.898684 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v76gx"] Oct 02 09:11:00 crc kubenswrapper[4960]: I1002 09:11:00.902920 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:00 crc kubenswrapper[4960]: I1002 09:11:00.914169 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v76gx"] Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.017249 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thgkc\" (UniqueName: \"kubernetes.io/projected/fa1192eb-7242-4d3d-8022-2eacac4ed402-kube-api-access-thgkc\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.017308 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-catalog-content\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.017345 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-utilities\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.120019 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-catalog-content\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.120083 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-utilities\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.120301 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thgkc\" (UniqueName: \"kubernetes.io/projected/fa1192eb-7242-4d3d-8022-2eacac4ed402-kube-api-access-thgkc\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.121267 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-utilities\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.121431 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-catalog-content\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.144859 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thgkc\" (UniqueName: \"kubernetes.io/projected/fa1192eb-7242-4d3d-8022-2eacac4ed402-kube-api-access-thgkc\") pod \"redhat-marketplace-v76gx\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.264915 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.639518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9"} Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.688249 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.688355 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.741908 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v76gx"] Oct 02 09:11:01 crc kubenswrapper[4960]: I1002 09:11:01.758018 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:11:01 crc kubenswrapper[4960]: W1002 09:11:01.760567 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa1192eb_7242_4d3d_8022_2eacac4ed402.slice/crio-12a2fb4277853100163a80734517b67c0a7d3cca55dade484a4b5a511d74fb39 WatchSource:0}: Error finding container 12a2fb4277853100163a80734517b67c0a7d3cca55dade484a4b5a511d74fb39: Status 404 returned error can't find the container with id 12a2fb4277853100163a80734517b67c0a7d3cca55dade484a4b5a511d74fb39 Oct 02 09:11:02 crc kubenswrapper[4960]: I1002 09:11:02.664307 4960 generic.go:334] "Generic (PLEG): container finished" podID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerID="a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c" exitCode=0 Oct 02 09:11:02 crc kubenswrapper[4960]: I1002 09:11:02.665200 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v76gx" event={"ID":"fa1192eb-7242-4d3d-8022-2eacac4ed402","Type":"ContainerDied","Data":"a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c"} Oct 02 09:11:02 crc kubenswrapper[4960]: I1002 09:11:02.665448 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v76gx" event={"ID":"fa1192eb-7242-4d3d-8022-2eacac4ed402","Type":"ContainerStarted","Data":"12a2fb4277853100163a80734517b67c0a7d3cca55dade484a4b5a511d74fb39"} Oct 02 09:11:02 crc kubenswrapper[4960]: I1002 09:11:02.733836 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:11:04 crc kubenswrapper[4960]: I1002 09:11:04.058782 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xbcjd"] Oct 02 09:11:04 crc kubenswrapper[4960]: I1002 09:11:04.689063 4960 generic.go:334] "Generic (PLEG): container finished" podID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerID="b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2" exitCode=0 Oct 02 09:11:04 crc kubenswrapper[4960]: I1002 09:11:04.689129 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v76gx" event={"ID":"fa1192eb-7242-4d3d-8022-2eacac4ed402","Type":"ContainerDied","Data":"b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2"} Oct 02 09:11:04 crc kubenswrapper[4960]: I1002 09:11:04.689338 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xbcjd" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="registry-server" containerID="cri-o://d96cc644bd52cdfa7b575afaf735b201dc58be52627009f86589430b25422bf6" gracePeriod=2 Oct 02 09:11:04 crc kubenswrapper[4960]: E1002 09:11:04.970373 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb805a38d_5112_4a9d_b30b_59635596aacc.slice/crio-d96cc644bd52cdfa7b575afaf735b201dc58be52627009f86589430b25422bf6.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.701553 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v76gx" event={"ID":"fa1192eb-7242-4d3d-8022-2eacac4ed402","Type":"ContainerStarted","Data":"a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32"} Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.706901 4960 generic.go:334] "Generic (PLEG): container finished" podID="b805a38d-5112-4a9d-b30b-59635596aacc" containerID="d96cc644bd52cdfa7b575afaf735b201dc58be52627009f86589430b25422bf6" exitCode=0 Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.706936 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbcjd" event={"ID":"b805a38d-5112-4a9d-b30b-59635596aacc","Type":"ContainerDied","Data":"d96cc644bd52cdfa7b575afaf735b201dc58be52627009f86589430b25422bf6"} Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.730817 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v76gx" podStartSLOduration=3.110710016 podStartE2EDuration="5.730790905s" podCreationTimestamp="2025-10-02 09:11:00 +0000 UTC" firstStartedPulling="2025-10-02 09:11:02.670997957 +0000 UTC m=+6883.702944244" lastFinishedPulling="2025-10-02 09:11:05.291078826 +0000 UTC m=+6886.323025133" observedRunningTime="2025-10-02 09:11:05.729062849 +0000 UTC m=+6886.761009146" watchObservedRunningTime="2025-10-02 09:11:05.730790905 +0000 UTC m=+6886.762737202" Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.843824 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.877630 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4zmg\" (UniqueName: \"kubernetes.io/projected/b805a38d-5112-4a9d-b30b-59635596aacc-kube-api-access-q4zmg\") pod \"b805a38d-5112-4a9d-b30b-59635596aacc\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.877893 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-utilities\") pod \"b805a38d-5112-4a9d-b30b-59635596aacc\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.878124 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-catalog-content\") pod \"b805a38d-5112-4a9d-b30b-59635596aacc\" (UID: \"b805a38d-5112-4a9d-b30b-59635596aacc\") " Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.885023 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-utilities" (OuterVolumeSpecName: "utilities") pod "b805a38d-5112-4a9d-b30b-59635596aacc" (UID: "b805a38d-5112-4a9d-b30b-59635596aacc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.915163 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b805a38d-5112-4a9d-b30b-59635596aacc-kube-api-access-q4zmg" (OuterVolumeSpecName: "kube-api-access-q4zmg") pod "b805a38d-5112-4a9d-b30b-59635596aacc" (UID: "b805a38d-5112-4a9d-b30b-59635596aacc"). InnerVolumeSpecName "kube-api-access-q4zmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.982347 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4zmg\" (UniqueName: \"kubernetes.io/projected/b805a38d-5112-4a9d-b30b-59635596aacc-kube-api-access-q4zmg\") on node \"crc\" DevicePath \"\"" Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.983355 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:11:05 crc kubenswrapper[4960]: I1002 09:11:05.983391 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b805a38d-5112-4a9d-b30b-59635596aacc" (UID: "b805a38d-5112-4a9d-b30b-59635596aacc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.086624 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b805a38d-5112-4a9d-b30b-59635596aacc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.736225 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbcjd" event={"ID":"b805a38d-5112-4a9d-b30b-59635596aacc","Type":"ContainerDied","Data":"9e0c7150188366a87358898a521a4c9bf893f0557bcb6ed1e18fdacfd895b001"} Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.736287 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbcjd" Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.736303 4960 scope.go:117] "RemoveContainer" containerID="d96cc644bd52cdfa7b575afaf735b201dc58be52627009f86589430b25422bf6" Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.777381 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xbcjd"] Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.783752 4960 scope.go:117] "RemoveContainer" containerID="8c14243de13686db01f255221b09b20370e7c90a59a0977a7b7b506ef6887f45" Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.786077 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xbcjd"] Oct 02 09:11:06 crc kubenswrapper[4960]: I1002 09:11:06.810743 4960 scope.go:117] "RemoveContainer" containerID="dec8e44a4ac5d90fe731d809995a0133ffe0690a5156169cd888566ce5760251" Oct 02 09:11:08 crc kubenswrapper[4960]: I1002 09:11:08.342624 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" path="/var/lib/kubelet/pods/b805a38d-5112-4a9d-b30b-59635596aacc/volumes" Oct 02 09:11:11 crc kubenswrapper[4960]: I1002 09:11:11.265911 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:11 crc kubenswrapper[4960]: I1002 09:11:11.266609 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:11 crc kubenswrapper[4960]: I1002 09:11:11.326328 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:11 crc kubenswrapper[4960]: I1002 09:11:11.887427 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:11 crc kubenswrapper[4960]: I1002 09:11:11.952065 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v76gx"] Oct 02 09:11:13 crc kubenswrapper[4960]: I1002 09:11:13.846057 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v76gx" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="registry-server" containerID="cri-o://a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32" gracePeriod=2 Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.416092 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.505369 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-utilities\") pod \"fa1192eb-7242-4d3d-8022-2eacac4ed402\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.506042 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thgkc\" (UniqueName: \"kubernetes.io/projected/fa1192eb-7242-4d3d-8022-2eacac4ed402-kube-api-access-thgkc\") pod \"fa1192eb-7242-4d3d-8022-2eacac4ed402\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.506217 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-catalog-content\") pod \"fa1192eb-7242-4d3d-8022-2eacac4ed402\" (UID: \"fa1192eb-7242-4d3d-8022-2eacac4ed402\") " Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.507178 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-utilities" (OuterVolumeSpecName: "utilities") pod "fa1192eb-7242-4d3d-8022-2eacac4ed402" (UID: "fa1192eb-7242-4d3d-8022-2eacac4ed402"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.514310 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa1192eb-7242-4d3d-8022-2eacac4ed402-kube-api-access-thgkc" (OuterVolumeSpecName: "kube-api-access-thgkc") pod "fa1192eb-7242-4d3d-8022-2eacac4ed402" (UID: "fa1192eb-7242-4d3d-8022-2eacac4ed402"). InnerVolumeSpecName "kube-api-access-thgkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.520925 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa1192eb-7242-4d3d-8022-2eacac4ed402" (UID: "fa1192eb-7242-4d3d-8022-2eacac4ed402"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.609817 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.609874 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1192eb-7242-4d3d-8022-2eacac4ed402-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.609889 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thgkc\" (UniqueName: \"kubernetes.io/projected/fa1192eb-7242-4d3d-8022-2eacac4ed402-kube-api-access-thgkc\") on node \"crc\" DevicePath \"\"" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.860802 4960 generic.go:334] "Generic (PLEG): container finished" podID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerID="a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32" exitCode=0 Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.860860 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v76gx" event={"ID":"fa1192eb-7242-4d3d-8022-2eacac4ed402","Type":"ContainerDied","Data":"a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32"} Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.860919 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v76gx" event={"ID":"fa1192eb-7242-4d3d-8022-2eacac4ed402","Type":"ContainerDied","Data":"12a2fb4277853100163a80734517b67c0a7d3cca55dade484a4b5a511d74fb39"} Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.860938 4960 scope.go:117] "RemoveContainer" containerID="a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.860943 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v76gx" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.886125 4960 scope.go:117] "RemoveContainer" containerID="b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.919301 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v76gx"] Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.924224 4960 scope.go:117] "RemoveContainer" containerID="a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.930021 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v76gx"] Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.988770 4960 scope.go:117] "RemoveContainer" containerID="a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32" Oct 02 09:11:14 crc kubenswrapper[4960]: E1002 09:11:14.990388 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32\": container with ID starting with a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32 not found: ID does not exist" containerID="a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.990483 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32"} err="failed to get container status \"a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32\": rpc error: code = NotFound desc = could not find container \"a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32\": container with ID starting with a7254aefcbe8761650b7e762a8f5c60a8fb5a74369583edef4d40ce3c4f23c32 not found: ID does not exist" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.990537 4960 scope.go:117] "RemoveContainer" containerID="b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2" Oct 02 09:11:14 crc kubenswrapper[4960]: E1002 09:11:14.991196 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2\": container with ID starting with b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2 not found: ID does not exist" containerID="b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.991242 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2"} err="failed to get container status \"b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2\": rpc error: code = NotFound desc = could not find container \"b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2\": container with ID starting with b3f34031fb1144851833319f7dfe463f052236fc93e3fc3fd20033f9ad705af2 not found: ID does not exist" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.991276 4960 scope.go:117] "RemoveContainer" containerID="a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c" Oct 02 09:11:14 crc kubenswrapper[4960]: E1002 09:11:14.992059 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c\": container with ID starting with a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c not found: ID does not exist" containerID="a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c" Oct 02 09:11:14 crc kubenswrapper[4960]: I1002 09:11:14.992172 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c"} err="failed to get container status \"a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c\": rpc error: code = NotFound desc = could not find container \"a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c\": container with ID starting with a539a7046e011283a568d4866d61e24a9b7807ab71e5548afb2279eae21d438c not found: ID does not exist" Oct 02 09:11:16 crc kubenswrapper[4960]: I1002 09:11:16.344817 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" path="/var/lib/kubelet/pods/fa1192eb-7242-4d3d-8022-2eacac4ed402/volumes" Oct 02 09:13:29 crc kubenswrapper[4960]: I1002 09:13:29.149971 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:13:29 crc kubenswrapper[4960]: I1002 09:13:29.150882 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:13:59 crc kubenswrapper[4960]: I1002 09:13:59.150686 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:13:59 crc kubenswrapper[4960]: I1002 09:13:59.151653 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.150461 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.151257 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.151313 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.152234 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.152291 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" gracePeriod=600 Oct 02 09:14:29 crc kubenswrapper[4960]: E1002 09:14:29.277553 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.946839 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" exitCode=0 Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.946901 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9"} Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.946955 4960 scope.go:117] "RemoveContainer" containerID="24fa21f7e874bc070325e0ed12f94c2337c713653061abc6301926e6dd32a6f6" Oct 02 09:14:29 crc kubenswrapper[4960]: I1002 09:14:29.949615 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:14:29 crc kubenswrapper[4960]: E1002 09:14:29.950290 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:14:41 crc kubenswrapper[4960]: I1002 09:14:41.330563 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:14:41 crc kubenswrapper[4960]: E1002 09:14:41.332425 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:14:55 crc kubenswrapper[4960]: I1002 09:14:55.331956 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:14:55 crc kubenswrapper[4960]: E1002 09:14:55.333321 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.194830 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn"] Oct 02 09:15:00 crc kubenswrapper[4960]: E1002 09:15:00.195860 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="registry-server" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.195878 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="registry-server" Oct 02 09:15:00 crc kubenswrapper[4960]: E1002 09:15:00.195911 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="extract-content" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.195922 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="extract-content" Oct 02 09:15:00 crc kubenswrapper[4960]: E1002 09:15:00.195935 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="extract-content" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.195946 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="extract-content" Oct 02 09:15:00 crc kubenswrapper[4960]: E1002 09:15:00.195958 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="extract-utilities" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.195968 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="extract-utilities" Oct 02 09:15:00 crc kubenswrapper[4960]: E1002 09:15:00.196013 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="extract-utilities" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.196023 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="extract-utilities" Oct 02 09:15:00 crc kubenswrapper[4960]: E1002 09:15:00.196051 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="registry-server" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.196059 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="registry-server" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.196325 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa1192eb-7242-4d3d-8022-2eacac4ed402" containerName="registry-server" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.196358 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b805a38d-5112-4a9d-b30b-59635596aacc" containerName="registry-server" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.197330 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.201146 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.204105 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.212574 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn"] Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.247924 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tjnv\" (UniqueName: \"kubernetes.io/projected/24a96346-3d6a-4a46-a0d4-66050b2019cf-kube-api-access-7tjnv\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.248816 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24a96346-3d6a-4a46-a0d4-66050b2019cf-config-volume\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.248911 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24a96346-3d6a-4a46-a0d4-66050b2019cf-secret-volume\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.358143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tjnv\" (UniqueName: \"kubernetes.io/projected/24a96346-3d6a-4a46-a0d4-66050b2019cf-kube-api-access-7tjnv\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.358424 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24a96346-3d6a-4a46-a0d4-66050b2019cf-config-volume\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.358522 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24a96346-3d6a-4a46-a0d4-66050b2019cf-secret-volume\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.360341 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24a96346-3d6a-4a46-a0d4-66050b2019cf-config-volume\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.379990 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24a96346-3d6a-4a46-a0d4-66050b2019cf-secret-volume\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.385228 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tjnv\" (UniqueName: \"kubernetes.io/projected/24a96346-3d6a-4a46-a0d4-66050b2019cf-kube-api-access-7tjnv\") pod \"collect-profiles-29323275-f7znn\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.522819 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:00 crc kubenswrapper[4960]: I1002 09:15:00.996502 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn"] Oct 02 09:15:01 crc kubenswrapper[4960]: I1002 09:15:01.326103 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" event={"ID":"24a96346-3d6a-4a46-a0d4-66050b2019cf","Type":"ContainerStarted","Data":"50b34e471c46a3249729afc4b12480843efb746c18b6669a51931bdd15bccea7"} Oct 02 09:15:01 crc kubenswrapper[4960]: I1002 09:15:01.326713 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" event={"ID":"24a96346-3d6a-4a46-a0d4-66050b2019cf","Type":"ContainerStarted","Data":"69ac58ff616462e60977fa01d768cbc0297970a2a8d520efb15a83c48cc52dd3"} Oct 02 09:15:01 crc kubenswrapper[4960]: I1002 09:15:01.346763 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" podStartSLOduration=1.34673473 podStartE2EDuration="1.34673473s" podCreationTimestamp="2025-10-02 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:15:01.344022447 +0000 UTC m=+7122.375968744" watchObservedRunningTime="2025-10-02 09:15:01.34673473 +0000 UTC m=+7122.378681027" Oct 02 09:15:02 crc kubenswrapper[4960]: I1002 09:15:02.338626 4960 generic.go:334] "Generic (PLEG): container finished" podID="24a96346-3d6a-4a46-a0d4-66050b2019cf" containerID="50b34e471c46a3249729afc4b12480843efb746c18b6669a51931bdd15bccea7" exitCode=0 Oct 02 09:15:02 crc kubenswrapper[4960]: I1002 09:15:02.351855 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" event={"ID":"24a96346-3d6a-4a46-a0d4-66050b2019cf","Type":"ContainerDied","Data":"50b34e471c46a3249729afc4b12480843efb746c18b6669a51931bdd15bccea7"} Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.793789 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.842219 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24a96346-3d6a-4a46-a0d4-66050b2019cf-secret-volume\") pod \"24a96346-3d6a-4a46-a0d4-66050b2019cf\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.842399 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24a96346-3d6a-4a46-a0d4-66050b2019cf-config-volume\") pod \"24a96346-3d6a-4a46-a0d4-66050b2019cf\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.842496 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tjnv\" (UniqueName: \"kubernetes.io/projected/24a96346-3d6a-4a46-a0d4-66050b2019cf-kube-api-access-7tjnv\") pod \"24a96346-3d6a-4a46-a0d4-66050b2019cf\" (UID: \"24a96346-3d6a-4a46-a0d4-66050b2019cf\") " Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.843471 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24a96346-3d6a-4a46-a0d4-66050b2019cf-config-volume" (OuterVolumeSpecName: "config-volume") pod "24a96346-3d6a-4a46-a0d4-66050b2019cf" (UID: "24a96346-3d6a-4a46-a0d4-66050b2019cf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.852135 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24a96346-3d6a-4a46-a0d4-66050b2019cf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "24a96346-3d6a-4a46-a0d4-66050b2019cf" (UID: "24a96346-3d6a-4a46-a0d4-66050b2019cf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.853270 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24a96346-3d6a-4a46-a0d4-66050b2019cf-kube-api-access-7tjnv" (OuterVolumeSpecName: "kube-api-access-7tjnv") pod "24a96346-3d6a-4a46-a0d4-66050b2019cf" (UID: "24a96346-3d6a-4a46-a0d4-66050b2019cf"). InnerVolumeSpecName "kube-api-access-7tjnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.944913 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24a96346-3d6a-4a46-a0d4-66050b2019cf-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.944951 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24a96346-3d6a-4a46-a0d4-66050b2019cf-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:15:03 crc kubenswrapper[4960]: I1002 09:15:03.945050 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tjnv\" (UniqueName: \"kubernetes.io/projected/24a96346-3d6a-4a46-a0d4-66050b2019cf-kube-api-access-7tjnv\") on node \"crc\" DevicePath \"\"" Oct 02 09:15:04 crc kubenswrapper[4960]: I1002 09:15:04.374313 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" event={"ID":"24a96346-3d6a-4a46-a0d4-66050b2019cf","Type":"ContainerDied","Data":"69ac58ff616462e60977fa01d768cbc0297970a2a8d520efb15a83c48cc52dd3"} Oct 02 09:15:04 crc kubenswrapper[4960]: I1002 09:15:04.374863 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69ac58ff616462e60977fa01d768cbc0297970a2a8d520efb15a83c48cc52dd3" Oct 02 09:15:04 crc kubenswrapper[4960]: I1002 09:15:04.374415 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn" Oct 02 09:15:04 crc kubenswrapper[4960]: I1002 09:15:04.441147 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv"] Oct 02 09:15:04 crc kubenswrapper[4960]: I1002 09:15:04.452961 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-n45xv"] Oct 02 09:15:06 crc kubenswrapper[4960]: I1002 09:15:06.346576 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d11f2c48-b7f7-4106-aedc-e2b94cab2798" path="/var/lib/kubelet/pods/d11f2c48-b7f7-4106-aedc-e2b94cab2798/volumes" Oct 02 09:15:08 crc kubenswrapper[4960]: I1002 09:15:08.330738 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:15:08 crc kubenswrapper[4960]: E1002 09:15:08.331850 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:15:23 crc kubenswrapper[4960]: I1002 09:15:23.330785 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:15:23 crc kubenswrapper[4960]: E1002 09:15:23.331638 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:15:36 crc kubenswrapper[4960]: I1002 09:15:36.331368 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:15:36 crc kubenswrapper[4960]: E1002 09:15:36.332780 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:15:48 crc kubenswrapper[4960]: I1002 09:15:48.330349 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:15:48 crc kubenswrapper[4960]: E1002 09:15:48.331696 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:16:03 crc kubenswrapper[4960]: I1002 09:16:03.330765 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:16:03 crc kubenswrapper[4960]: E1002 09:16:03.332237 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:16:04 crc kubenswrapper[4960]: I1002 09:16:04.372138 4960 scope.go:117] "RemoveContainer" containerID="0d7ac2e1c12432571509cdd04e63c9c47b4e2c4ef0a546fae51861ecbec23b68" Oct 02 09:16:18 crc kubenswrapper[4960]: I1002 09:16:18.330512 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:16:18 crc kubenswrapper[4960]: E1002 09:16:18.331439 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:16:31 crc kubenswrapper[4960]: I1002 09:16:31.330146 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:16:31 crc kubenswrapper[4960]: E1002 09:16:31.331055 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:16:46 crc kubenswrapper[4960]: I1002 09:16:46.330416 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:16:46 crc kubenswrapper[4960]: E1002 09:16:46.331339 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:16:59 crc kubenswrapper[4960]: I1002 09:16:59.330029 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:16:59 crc kubenswrapper[4960]: E1002 09:16:59.331020 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:17:14 crc kubenswrapper[4960]: I1002 09:17:14.330094 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:17:14 crc kubenswrapper[4960]: E1002 09:17:14.331263 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:17:25 crc kubenswrapper[4960]: I1002 09:17:25.330994 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:17:25 crc kubenswrapper[4960]: E1002 09:17:25.332927 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:17:39 crc kubenswrapper[4960]: I1002 09:17:39.330360 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:17:39 crc kubenswrapper[4960]: E1002 09:17:39.332292 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:17:50 crc kubenswrapper[4960]: I1002 09:17:50.337524 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:17:50 crc kubenswrapper[4960]: E1002 09:17:50.339512 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:18:02 crc kubenswrapper[4960]: I1002 09:18:02.330252 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:18:02 crc kubenswrapper[4960]: E1002 09:18:02.331301 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:18:14 crc kubenswrapper[4960]: I1002 09:18:14.329870 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:18:14 crc kubenswrapper[4960]: E1002 09:18:14.330724 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:18:28 crc kubenswrapper[4960]: I1002 09:18:28.330743 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:18:28 crc kubenswrapper[4960]: E1002 09:18:28.331837 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:18:42 crc kubenswrapper[4960]: I1002 09:18:42.335255 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:18:42 crc kubenswrapper[4960]: E1002 09:18:42.336342 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:18:56 crc kubenswrapper[4960]: I1002 09:18:56.330442 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:18:56 crc kubenswrapper[4960]: E1002 09:18:56.331616 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:19:09 crc kubenswrapper[4960]: I1002 09:19:09.330617 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:19:09 crc kubenswrapper[4960]: E1002 09:19:09.332033 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:19:20 crc kubenswrapper[4960]: I1002 09:19:20.343249 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:19:20 crc kubenswrapper[4960]: E1002 09:19:20.344992 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:19:32 crc kubenswrapper[4960]: I1002 09:19:32.330878 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:19:33 crc kubenswrapper[4960]: I1002 09:19:33.125336 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"e25a36f3fbf7aec29094839309be354cbf21ad21e0ffe21d97b0d57f27fae0b5"} Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.814612 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z4pfq"] Oct 02 09:21:11 crc kubenswrapper[4960]: E1002 09:21:11.815538 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a96346-3d6a-4a46-a0d4-66050b2019cf" containerName="collect-profiles" Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.815551 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a96346-3d6a-4a46-a0d4-66050b2019cf" containerName="collect-profiles" Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.815713 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="24a96346-3d6a-4a46-a0d4-66050b2019cf" containerName="collect-profiles" Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.817439 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.842572 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z4pfq"] Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.897310 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-utilities\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.897948 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-catalog-content\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:11 crc kubenswrapper[4960]: I1002 09:21:11.898035 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wzpm\" (UniqueName: \"kubernetes.io/projected/d1658398-a41e-460d-92cc-810409eadfb9-kube-api-access-5wzpm\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.001285 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-utilities\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.001468 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-catalog-content\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.001505 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wzpm\" (UniqueName: \"kubernetes.io/projected/d1658398-a41e-460d-92cc-810409eadfb9-kube-api-access-5wzpm\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.002071 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-utilities\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.002096 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-catalog-content\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.024870 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wzpm\" (UniqueName: \"kubernetes.io/projected/d1658398-a41e-460d-92cc-810409eadfb9-kube-api-access-5wzpm\") pod \"community-operators-z4pfq\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.144223 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:12 crc kubenswrapper[4960]: I1002 09:21:12.711527 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z4pfq"] Oct 02 09:21:13 crc kubenswrapper[4960]: I1002 09:21:13.198437 4960 generic.go:334] "Generic (PLEG): container finished" podID="d1658398-a41e-460d-92cc-810409eadfb9" containerID="5cf319829ea342951008c7353159405a4a1b19a4dc8f6a90c80d1167e4fa4ec4" exitCode=0 Oct 02 09:21:13 crc kubenswrapper[4960]: I1002 09:21:13.198514 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4pfq" event={"ID":"d1658398-a41e-460d-92cc-810409eadfb9","Type":"ContainerDied","Data":"5cf319829ea342951008c7353159405a4a1b19a4dc8f6a90c80d1167e4fa4ec4"} Oct 02 09:21:13 crc kubenswrapper[4960]: I1002 09:21:13.198848 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4pfq" event={"ID":"d1658398-a41e-460d-92cc-810409eadfb9","Type":"ContainerStarted","Data":"710c7c4681112d8d0639de1ef8dca77906ee168d21c82fcd39fa3f27f3aa4647"} Oct 02 09:21:13 crc kubenswrapper[4960]: I1002 09:21:13.203987 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:21:15 crc kubenswrapper[4960]: I1002 09:21:15.220441 4960 generic.go:334] "Generic (PLEG): container finished" podID="d1658398-a41e-460d-92cc-810409eadfb9" containerID="b3a7f5b03c99328152a53780633ceb2afc21f26a3a100af2f26b03cf251ae49b" exitCode=0 Oct 02 09:21:15 crc kubenswrapper[4960]: I1002 09:21:15.220508 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4pfq" event={"ID":"d1658398-a41e-460d-92cc-810409eadfb9","Type":"ContainerDied","Data":"b3a7f5b03c99328152a53780633ceb2afc21f26a3a100af2f26b03cf251ae49b"} Oct 02 09:21:17 crc kubenswrapper[4960]: I1002 09:21:17.244674 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4pfq" event={"ID":"d1658398-a41e-460d-92cc-810409eadfb9","Type":"ContainerStarted","Data":"711791d1161ba35fe0b5422f2375d0653b3fa6e94c073f7fc7100ea88b431b98"} Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.182592 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z4pfq" podStartSLOduration=3.9796454519999998 podStartE2EDuration="7.182568733s" podCreationTimestamp="2025-10-02 09:21:11 +0000 UTC" firstStartedPulling="2025-10-02 09:21:13.20369278 +0000 UTC m=+7494.235639067" lastFinishedPulling="2025-10-02 09:21:16.406616061 +0000 UTC m=+7497.438562348" observedRunningTime="2025-10-02 09:21:17.273141147 +0000 UTC m=+7498.305087444" watchObservedRunningTime="2025-10-02 09:21:18.182568733 +0000 UTC m=+7499.214515030" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.192512 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hkrhl"] Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.195141 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.205335 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hkrhl"] Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.255589 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-catalog-content\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.255722 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lmfr\" (UniqueName: \"kubernetes.io/projected/669164a8-eb82-424e-9a2e-014ac84bca5f-kube-api-access-8lmfr\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.255776 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-utilities\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.357910 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-catalog-content\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.358054 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lmfr\" (UniqueName: \"kubernetes.io/projected/669164a8-eb82-424e-9a2e-014ac84bca5f-kube-api-access-8lmfr\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.358128 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-utilities\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.358705 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-catalog-content\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.358799 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-utilities\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.380272 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lmfr\" (UniqueName: \"kubernetes.io/projected/669164a8-eb82-424e-9a2e-014ac84bca5f-kube-api-access-8lmfr\") pod \"redhat-operators-hkrhl\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:18 crc kubenswrapper[4960]: I1002 09:21:18.515895 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:19 crc kubenswrapper[4960]: I1002 09:21:19.157088 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hkrhl"] Oct 02 09:21:19 crc kubenswrapper[4960]: I1002 09:21:19.268371 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hkrhl" event={"ID":"669164a8-eb82-424e-9a2e-014ac84bca5f","Type":"ContainerStarted","Data":"35d4a779b532d0476d526fedcc1d670b4eeaf6396602c0ae3d15876dcbab3ebd"} Oct 02 09:21:20 crc kubenswrapper[4960]: I1002 09:21:20.285266 4960 generic.go:334] "Generic (PLEG): container finished" podID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerID="26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db" exitCode=0 Oct 02 09:21:20 crc kubenswrapper[4960]: I1002 09:21:20.285364 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hkrhl" event={"ID":"669164a8-eb82-424e-9a2e-014ac84bca5f","Type":"ContainerDied","Data":"26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db"} Oct 02 09:21:22 crc kubenswrapper[4960]: I1002 09:21:22.144655 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:22 crc kubenswrapper[4960]: I1002 09:21:22.145437 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:22 crc kubenswrapper[4960]: I1002 09:21:22.207417 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:22 crc kubenswrapper[4960]: I1002 09:21:22.311551 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hkrhl" event={"ID":"669164a8-eb82-424e-9a2e-014ac84bca5f","Type":"ContainerStarted","Data":"2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85"} Oct 02 09:21:22 crc kubenswrapper[4960]: I1002 09:21:22.374528 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:24 crc kubenswrapper[4960]: I1002 09:21:23.323248 4960 generic.go:334] "Generic (PLEG): container finished" podID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerID="2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85" exitCode=0 Oct 02 09:21:24 crc kubenswrapper[4960]: I1002 09:21:23.323334 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hkrhl" event={"ID":"669164a8-eb82-424e-9a2e-014ac84bca5f","Type":"ContainerDied","Data":"2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85"} Oct 02 09:21:24 crc kubenswrapper[4960]: I1002 09:21:24.790638 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z4pfq"] Oct 02 09:21:24 crc kubenswrapper[4960]: I1002 09:21:24.791443 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z4pfq" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="registry-server" containerID="cri-o://711791d1161ba35fe0b5422f2375d0653b3fa6e94c073f7fc7100ea88b431b98" gracePeriod=2 Oct 02 09:21:27 crc kubenswrapper[4960]: I1002 09:21:27.372145 4960 generic.go:334] "Generic (PLEG): container finished" podID="d1658398-a41e-460d-92cc-810409eadfb9" containerID="711791d1161ba35fe0b5422f2375d0653b3fa6e94c073f7fc7100ea88b431b98" exitCode=0 Oct 02 09:21:27 crc kubenswrapper[4960]: I1002 09:21:27.372219 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4pfq" event={"ID":"d1658398-a41e-460d-92cc-810409eadfb9","Type":"ContainerDied","Data":"711791d1161ba35fe0b5422f2375d0653b3fa6e94c073f7fc7100ea88b431b98"} Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.068250 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.114503 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-utilities\") pod \"d1658398-a41e-460d-92cc-810409eadfb9\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.115098 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wzpm\" (UniqueName: \"kubernetes.io/projected/d1658398-a41e-460d-92cc-810409eadfb9-kube-api-access-5wzpm\") pod \"d1658398-a41e-460d-92cc-810409eadfb9\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.115378 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-catalog-content\") pod \"d1658398-a41e-460d-92cc-810409eadfb9\" (UID: \"d1658398-a41e-460d-92cc-810409eadfb9\") " Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.117951 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-utilities" (OuterVolumeSpecName: "utilities") pod "d1658398-a41e-460d-92cc-810409eadfb9" (UID: "d1658398-a41e-460d-92cc-810409eadfb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.133451 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1658398-a41e-460d-92cc-810409eadfb9-kube-api-access-5wzpm" (OuterVolumeSpecName: "kube-api-access-5wzpm") pod "d1658398-a41e-460d-92cc-810409eadfb9" (UID: "d1658398-a41e-460d-92cc-810409eadfb9"). InnerVolumeSpecName "kube-api-access-5wzpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.175629 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1658398-a41e-460d-92cc-810409eadfb9" (UID: "d1658398-a41e-460d-92cc-810409eadfb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.219304 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.219352 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1658398-a41e-460d-92cc-810409eadfb9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.219364 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wzpm\" (UniqueName: \"kubernetes.io/projected/d1658398-a41e-460d-92cc-810409eadfb9-kube-api-access-5wzpm\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.390045 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4pfq" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.390428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4pfq" event={"ID":"d1658398-a41e-460d-92cc-810409eadfb9","Type":"ContainerDied","Data":"710c7c4681112d8d0639de1ef8dca77906ee168d21c82fcd39fa3f27f3aa4647"} Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.390504 4960 scope.go:117] "RemoveContainer" containerID="711791d1161ba35fe0b5422f2375d0653b3fa6e94c073f7fc7100ea88b431b98" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.400500 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hkrhl" event={"ID":"669164a8-eb82-424e-9a2e-014ac84bca5f","Type":"ContainerStarted","Data":"83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170"} Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.428089 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z4pfq"] Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.436459 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z4pfq"] Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.437333 4960 scope.go:117] "RemoveContainer" containerID="b3a7f5b03c99328152a53780633ceb2afc21f26a3a100af2f26b03cf251ae49b" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.447553 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hkrhl" podStartSLOduration=3.455711348 podStartE2EDuration="10.44752342s" podCreationTimestamp="2025-10-02 09:21:18 +0000 UTC" firstStartedPulling="2025-10-02 09:21:20.288937538 +0000 UTC m=+7501.320883825" lastFinishedPulling="2025-10-02 09:21:27.28074961 +0000 UTC m=+7508.312695897" observedRunningTime="2025-10-02 09:21:28.441226821 +0000 UTC m=+7509.473173108" watchObservedRunningTime="2025-10-02 09:21:28.44752342 +0000 UTC m=+7509.479469707" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.472111 4960 scope.go:117] "RemoveContainer" containerID="5cf319829ea342951008c7353159405a4a1b19a4dc8f6a90c80d1167e4fa4ec4" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.516718 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:28 crc kubenswrapper[4960]: I1002 09:21:28.516864 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:29 crc kubenswrapper[4960]: I1002 09:21:29.624001 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hkrhl" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="registry-server" probeResult="failure" output=< Oct 02 09:21:29 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 09:21:29 crc kubenswrapper[4960]: > Oct 02 09:21:30 crc kubenswrapper[4960]: I1002 09:21:30.412876 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1658398-a41e-460d-92cc-810409eadfb9" path="/var/lib/kubelet/pods/d1658398-a41e-460d-92cc-810409eadfb9/volumes" Oct 02 09:21:38 crc kubenswrapper[4960]: I1002 09:21:38.564752 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:38 crc kubenswrapper[4960]: I1002 09:21:38.624095 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:38 crc kubenswrapper[4960]: I1002 09:21:38.814155 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hkrhl"] Oct 02 09:21:40 crc kubenswrapper[4960]: I1002 09:21:40.525408 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hkrhl" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="registry-server" containerID="cri-o://83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170" gracePeriod=2 Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.126432 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.267538 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-catalog-content\") pod \"669164a8-eb82-424e-9a2e-014ac84bca5f\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.267808 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lmfr\" (UniqueName: \"kubernetes.io/projected/669164a8-eb82-424e-9a2e-014ac84bca5f-kube-api-access-8lmfr\") pod \"669164a8-eb82-424e-9a2e-014ac84bca5f\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.267996 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-utilities\") pod \"669164a8-eb82-424e-9a2e-014ac84bca5f\" (UID: \"669164a8-eb82-424e-9a2e-014ac84bca5f\") " Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.269007 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-utilities" (OuterVolumeSpecName: "utilities") pod "669164a8-eb82-424e-9a2e-014ac84bca5f" (UID: "669164a8-eb82-424e-9a2e-014ac84bca5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.279315 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/669164a8-eb82-424e-9a2e-014ac84bca5f-kube-api-access-8lmfr" (OuterVolumeSpecName: "kube-api-access-8lmfr") pod "669164a8-eb82-424e-9a2e-014ac84bca5f" (UID: "669164a8-eb82-424e-9a2e-014ac84bca5f"). InnerVolumeSpecName "kube-api-access-8lmfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.370689 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "669164a8-eb82-424e-9a2e-014ac84bca5f" (UID: "669164a8-eb82-424e-9a2e-014ac84bca5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.371038 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.371308 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lmfr\" (UniqueName: \"kubernetes.io/projected/669164a8-eb82-424e-9a2e-014ac84bca5f-kube-api-access-8lmfr\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.474331 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/669164a8-eb82-424e-9a2e-014ac84bca5f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.539270 4960 generic.go:334] "Generic (PLEG): container finished" podID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerID="83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170" exitCode=0 Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.539340 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hkrhl" event={"ID":"669164a8-eb82-424e-9a2e-014ac84bca5f","Type":"ContainerDied","Data":"83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170"} Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.539380 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hkrhl" event={"ID":"669164a8-eb82-424e-9a2e-014ac84bca5f","Type":"ContainerDied","Data":"35d4a779b532d0476d526fedcc1d670b4eeaf6396602c0ae3d15876dcbab3ebd"} Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.539409 4960 scope.go:117] "RemoveContainer" containerID="83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.539671 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hkrhl" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.582841 4960 scope.go:117] "RemoveContainer" containerID="2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.599847 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hkrhl"] Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.610923 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hkrhl"] Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.616125 4960 scope.go:117] "RemoveContainer" containerID="26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.671327 4960 scope.go:117] "RemoveContainer" containerID="83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170" Oct 02 09:21:41 crc kubenswrapper[4960]: E1002 09:21:41.671881 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170\": container with ID starting with 83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170 not found: ID does not exist" containerID="83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.671934 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170"} err="failed to get container status \"83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170\": rpc error: code = NotFound desc = could not find container \"83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170\": container with ID starting with 83f0933471d719f105f00e4ac7457afa064cd5b7d7749b85ec0bec9d399f5170 not found: ID does not exist" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.671969 4960 scope.go:117] "RemoveContainer" containerID="2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85" Oct 02 09:21:41 crc kubenswrapper[4960]: E1002 09:21:41.672522 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85\": container with ID starting with 2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85 not found: ID does not exist" containerID="2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.672558 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85"} err="failed to get container status \"2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85\": rpc error: code = NotFound desc = could not find container \"2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85\": container with ID starting with 2ec4a27c602c27658eef09724e310bf3f8a16cbca6212262ef76fe4f140a4b85 not found: ID does not exist" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.672586 4960 scope.go:117] "RemoveContainer" containerID="26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db" Oct 02 09:21:41 crc kubenswrapper[4960]: E1002 09:21:41.672853 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db\": container with ID starting with 26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db not found: ID does not exist" containerID="26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db" Oct 02 09:21:41 crc kubenswrapper[4960]: I1002 09:21:41.672879 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db"} err="failed to get container status \"26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db\": rpc error: code = NotFound desc = could not find container \"26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db\": container with ID starting with 26656098eff638fa8e6ad871c25f71e122771d975d2e2d05fb067b4d29df45db not found: ID does not exist" Oct 02 09:21:42 crc kubenswrapper[4960]: I1002 09:21:42.341850 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" path="/var/lib/kubelet/pods/669164a8-eb82-424e-9a2e-014ac84bca5f/volumes" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.223189 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4bksw"] Oct 02 09:21:45 crc kubenswrapper[4960]: E1002 09:21:45.224099 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="extract-utilities" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224115 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="extract-utilities" Oct 02 09:21:45 crc kubenswrapper[4960]: E1002 09:21:45.224144 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="extract-utilities" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224151 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="extract-utilities" Oct 02 09:21:45 crc kubenswrapper[4960]: E1002 09:21:45.224167 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="extract-content" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224174 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="extract-content" Oct 02 09:21:45 crc kubenswrapper[4960]: E1002 09:21:45.224189 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="registry-server" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224199 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="registry-server" Oct 02 09:21:45 crc kubenswrapper[4960]: E1002 09:21:45.224210 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="extract-content" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224217 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="extract-content" Oct 02 09:21:45 crc kubenswrapper[4960]: E1002 09:21:45.224263 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="registry-server" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224271 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="registry-server" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224482 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="669164a8-eb82-424e-9a2e-014ac84bca5f" containerName="registry-server" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.224590 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1658398-a41e-460d-92cc-810409eadfb9" containerName="registry-server" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.226606 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.261288 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bksw"] Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.382165 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-catalog-content\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.382481 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8fs2\" (UniqueName: \"kubernetes.io/projected/b65f6fd7-a5c0-4972-b544-d366898f3860-kube-api-access-c8fs2\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.382569 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-utilities\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.485499 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8fs2\" (UniqueName: \"kubernetes.io/projected/b65f6fd7-a5c0-4972-b544-d366898f3860-kube-api-access-c8fs2\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.485609 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-utilities\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.485875 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-catalog-content\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.486462 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-utilities\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.486551 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-catalog-content\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.520446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8fs2\" (UniqueName: \"kubernetes.io/projected/b65f6fd7-a5c0-4972-b544-d366898f3860-kube-api-access-c8fs2\") pod \"redhat-marketplace-4bksw\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:45 crc kubenswrapper[4960]: I1002 09:21:45.551524 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:46 crc kubenswrapper[4960]: I1002 09:21:46.084154 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bksw"] Oct 02 09:21:46 crc kubenswrapper[4960]: I1002 09:21:46.590833 4960 generic.go:334] "Generic (PLEG): container finished" podID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerID="61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f" exitCode=0 Oct 02 09:21:46 crc kubenswrapper[4960]: I1002 09:21:46.591022 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bksw" event={"ID":"b65f6fd7-a5c0-4972-b544-d366898f3860","Type":"ContainerDied","Data":"61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f"} Oct 02 09:21:46 crc kubenswrapper[4960]: I1002 09:21:46.591369 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bksw" event={"ID":"b65f6fd7-a5c0-4972-b544-d366898f3860","Type":"ContainerStarted","Data":"1485d6980fb52a6a2ee9d3a7ee88a78b9462175eb24dd4a5b5323cb977e13849"} Oct 02 09:21:48 crc kubenswrapper[4960]: I1002 09:21:48.616864 4960 generic.go:334] "Generic (PLEG): container finished" podID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerID="371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f" exitCode=0 Oct 02 09:21:48 crc kubenswrapper[4960]: I1002 09:21:48.617000 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bksw" event={"ID":"b65f6fd7-a5c0-4972-b544-d366898f3860","Type":"ContainerDied","Data":"371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f"} Oct 02 09:21:49 crc kubenswrapper[4960]: I1002 09:21:49.636349 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bksw" event={"ID":"b65f6fd7-a5c0-4972-b544-d366898f3860","Type":"ContainerStarted","Data":"de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee"} Oct 02 09:21:49 crc kubenswrapper[4960]: I1002 09:21:49.667356 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4bksw" podStartSLOduration=2.078328908 podStartE2EDuration="4.66733044s" podCreationTimestamp="2025-10-02 09:21:45 +0000 UTC" firstStartedPulling="2025-10-02 09:21:46.593200024 +0000 UTC m=+7527.625146321" lastFinishedPulling="2025-10-02 09:21:49.182201566 +0000 UTC m=+7530.214147853" observedRunningTime="2025-10-02 09:21:49.663501127 +0000 UTC m=+7530.695447414" watchObservedRunningTime="2025-10-02 09:21:49.66733044 +0000 UTC m=+7530.699276727" Oct 02 09:21:55 crc kubenswrapper[4960]: I1002 09:21:55.552045 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:55 crc kubenswrapper[4960]: I1002 09:21:55.553082 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:55 crc kubenswrapper[4960]: I1002 09:21:55.610837 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:55 crc kubenswrapper[4960]: I1002 09:21:55.765654 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:55 crc kubenswrapper[4960]: I1002 09:21:55.856659 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bksw"] Oct 02 09:21:57 crc kubenswrapper[4960]: I1002 09:21:57.739582 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4bksw" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="registry-server" containerID="cri-o://de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee" gracePeriod=2 Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.430233 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.629154 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-utilities\") pod \"b65f6fd7-a5c0-4972-b544-d366898f3860\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.629621 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8fs2\" (UniqueName: \"kubernetes.io/projected/b65f6fd7-a5c0-4972-b544-d366898f3860-kube-api-access-c8fs2\") pod \"b65f6fd7-a5c0-4972-b544-d366898f3860\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.630071 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-catalog-content\") pod \"b65f6fd7-a5c0-4972-b544-d366898f3860\" (UID: \"b65f6fd7-a5c0-4972-b544-d366898f3860\") " Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.630527 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-utilities" (OuterVolumeSpecName: "utilities") pod "b65f6fd7-a5c0-4972-b544-d366898f3860" (UID: "b65f6fd7-a5c0-4972-b544-d366898f3860"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.631176 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.640564 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b65f6fd7-a5c0-4972-b544-d366898f3860-kube-api-access-c8fs2" (OuterVolumeSpecName: "kube-api-access-c8fs2") pod "b65f6fd7-a5c0-4972-b544-d366898f3860" (UID: "b65f6fd7-a5c0-4972-b544-d366898f3860"). InnerVolumeSpecName "kube-api-access-c8fs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.645592 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b65f6fd7-a5c0-4972-b544-d366898f3860" (UID: "b65f6fd7-a5c0-4972-b544-d366898f3860"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.733836 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8fs2\" (UniqueName: \"kubernetes.io/projected/b65f6fd7-a5c0-4972-b544-d366898f3860-kube-api-access-c8fs2\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.733916 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65f6fd7-a5c0-4972-b544-d366898f3860-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.754034 4960 generic.go:334] "Generic (PLEG): container finished" podID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerID="de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee" exitCode=0 Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.754105 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bksw" event={"ID":"b65f6fd7-a5c0-4972-b544-d366898f3860","Type":"ContainerDied","Data":"de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee"} Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.754119 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bksw" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.754160 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bksw" event={"ID":"b65f6fd7-a5c0-4972-b544-d366898f3860","Type":"ContainerDied","Data":"1485d6980fb52a6a2ee9d3a7ee88a78b9462175eb24dd4a5b5323cb977e13849"} Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.754202 4960 scope.go:117] "RemoveContainer" containerID="de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.794825 4960 scope.go:117] "RemoveContainer" containerID="371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.798667 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bksw"] Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.822843 4960 scope.go:117] "RemoveContainer" containerID="61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.827776 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bksw"] Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.872712 4960 scope.go:117] "RemoveContainer" containerID="de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee" Oct 02 09:21:58 crc kubenswrapper[4960]: E1002 09:21:58.873325 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee\": container with ID starting with de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee not found: ID does not exist" containerID="de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.873397 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee"} err="failed to get container status \"de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee\": rpc error: code = NotFound desc = could not find container \"de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee\": container with ID starting with de4f19f5825ab7d6fd84406757f2aed6731e9f4f248e5e71ae06afe811dedbee not found: ID does not exist" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.873444 4960 scope.go:117] "RemoveContainer" containerID="371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f" Oct 02 09:21:58 crc kubenswrapper[4960]: E1002 09:21:58.874157 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f\": container with ID starting with 371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f not found: ID does not exist" containerID="371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.874213 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f"} err="failed to get container status \"371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f\": rpc error: code = NotFound desc = could not find container \"371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f\": container with ID starting with 371f06cb2a1589fe38b3e1f5adb9b78a3653949475312f793b599aa32add6d5f not found: ID does not exist" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.874258 4960 scope.go:117] "RemoveContainer" containerID="61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f" Oct 02 09:21:58 crc kubenswrapper[4960]: E1002 09:21:58.874691 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f\": container with ID starting with 61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f not found: ID does not exist" containerID="61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f" Oct 02 09:21:58 crc kubenswrapper[4960]: I1002 09:21:58.874759 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f"} err="failed to get container status \"61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f\": rpc error: code = NotFound desc = could not find container \"61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f\": container with ID starting with 61d4bfee9e6d5644b161f161ca98a6ed8483f9c3f89d7f5122a3a7c27f2d107f not found: ID does not exist" Oct 02 09:21:59 crc kubenswrapper[4960]: I1002 09:21:59.150348 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:21:59 crc kubenswrapper[4960]: I1002 09:21:59.150459 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:22:00 crc kubenswrapper[4960]: I1002 09:22:00.343144 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" path="/var/lib/kubelet/pods/b65f6fd7-a5c0-4972-b544-d366898f3860/volumes" Oct 02 09:22:29 crc kubenswrapper[4960]: I1002 09:22:29.150501 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:22:29 crc kubenswrapper[4960]: I1002 09:22:29.151311 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.150581 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.151494 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.151575 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.152758 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e25a36f3fbf7aec29094839309be354cbf21ad21e0ffe21d97b0d57f27fae0b5"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.152841 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://e25a36f3fbf7aec29094839309be354cbf21ad21e0ffe21d97b0d57f27fae0b5" gracePeriod=600 Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.408779 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="e25a36f3fbf7aec29094839309be354cbf21ad21e0ffe21d97b0d57f27fae0b5" exitCode=0 Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.408833 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"e25a36f3fbf7aec29094839309be354cbf21ad21e0ffe21d97b0d57f27fae0b5"} Oct 02 09:22:59 crc kubenswrapper[4960]: I1002 09:22:59.408877 4960 scope.go:117] "RemoveContainer" containerID="7522e81929334d33b50f08e55e79b83a89a43fc4e3a086ddc5ff77d600e277a9" Oct 02 09:23:00 crc kubenswrapper[4960]: I1002 09:23:00.428038 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7"} Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.662081 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zz9hm"] Oct 02 09:23:54 crc kubenswrapper[4960]: E1002 09:23:54.663653 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="extract-content" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.663673 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="extract-content" Oct 02 09:23:54 crc kubenswrapper[4960]: E1002 09:23:54.663717 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="extract-utilities" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.663726 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="extract-utilities" Oct 02 09:23:54 crc kubenswrapper[4960]: E1002 09:23:54.663740 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="registry-server" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.663747 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="registry-server" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.664003 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b65f6fd7-a5c0-4972-b544-d366898f3860" containerName="registry-server" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.668083 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.690860 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zz9hm"] Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.785405 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kff4v\" (UniqueName: \"kubernetes.io/projected/4ea495fa-92fb-463e-a12e-2a134f84b0ea-kube-api-access-kff4v\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.785552 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-utilities\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.785873 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-catalog-content\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.888323 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-catalog-content\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.888431 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kff4v\" (UniqueName: \"kubernetes.io/projected/4ea495fa-92fb-463e-a12e-2a134f84b0ea-kube-api-access-kff4v\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.888472 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-utilities\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.889069 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-catalog-content\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.889194 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-utilities\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:54 crc kubenswrapper[4960]: I1002 09:23:54.932657 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kff4v\" (UniqueName: \"kubernetes.io/projected/4ea495fa-92fb-463e-a12e-2a134f84b0ea-kube-api-access-kff4v\") pod \"certified-operators-zz9hm\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:55 crc kubenswrapper[4960]: I1002 09:23:55.009652 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:23:55 crc kubenswrapper[4960]: I1002 09:23:55.462407 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zz9hm"] Oct 02 09:23:56 crc kubenswrapper[4960]: I1002 09:23:56.042365 4960 generic.go:334] "Generic (PLEG): container finished" podID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerID="02d243db984973a4108e203c544f86c856842f0f6d0a40b65fde6b347ccb444a" exitCode=0 Oct 02 09:23:56 crc kubenswrapper[4960]: I1002 09:23:56.042422 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zz9hm" event={"ID":"4ea495fa-92fb-463e-a12e-2a134f84b0ea","Type":"ContainerDied","Data":"02d243db984973a4108e203c544f86c856842f0f6d0a40b65fde6b347ccb444a"} Oct 02 09:23:56 crc kubenswrapper[4960]: I1002 09:23:56.042927 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zz9hm" event={"ID":"4ea495fa-92fb-463e-a12e-2a134f84b0ea","Type":"ContainerStarted","Data":"d1a2005d9b19bcf0edec697a96bdd7f058fb7d25e01582d92afe7b8a150df252"} Oct 02 09:24:00 crc kubenswrapper[4960]: I1002 09:24:00.106259 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zz9hm" event={"ID":"4ea495fa-92fb-463e-a12e-2a134f84b0ea","Type":"ContainerStarted","Data":"bd764d472f75fdab75af95ca54ad6100afdafd5a83cd48a809b50b86a17bccee"} Oct 02 09:24:04 crc kubenswrapper[4960]: I1002 09:24:04.144196 4960 generic.go:334] "Generic (PLEG): container finished" podID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerID="bd764d472f75fdab75af95ca54ad6100afdafd5a83cd48a809b50b86a17bccee" exitCode=0 Oct 02 09:24:04 crc kubenswrapper[4960]: I1002 09:24:04.144314 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zz9hm" event={"ID":"4ea495fa-92fb-463e-a12e-2a134f84b0ea","Type":"ContainerDied","Data":"bd764d472f75fdab75af95ca54ad6100afdafd5a83cd48a809b50b86a17bccee"} Oct 02 09:24:06 crc kubenswrapper[4960]: I1002 09:24:06.171899 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zz9hm" event={"ID":"4ea495fa-92fb-463e-a12e-2a134f84b0ea","Type":"ContainerStarted","Data":"ea376c0dfa01457872c0504ae352c3cd4f665564ea4e0f440135c830de648c5a"} Oct 02 09:24:06 crc kubenswrapper[4960]: I1002 09:24:06.211867 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zz9hm" podStartSLOduration=2.91890225 podStartE2EDuration="12.211828241s" podCreationTimestamp="2025-10-02 09:23:54 +0000 UTC" firstStartedPulling="2025-10-02 09:23:56.045295655 +0000 UTC m=+7657.077241942" lastFinishedPulling="2025-10-02 09:24:05.338221626 +0000 UTC m=+7666.370167933" observedRunningTime="2025-10-02 09:24:06.196725167 +0000 UTC m=+7667.228671454" watchObservedRunningTime="2025-10-02 09:24:06.211828241 +0000 UTC m=+7667.243774528" Oct 02 09:24:15 crc kubenswrapper[4960]: I1002 09:24:15.010444 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:24:15 crc kubenswrapper[4960]: I1002 09:24:15.011257 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:24:15 crc kubenswrapper[4960]: I1002 09:24:15.061870 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:24:15 crc kubenswrapper[4960]: I1002 09:24:15.320840 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:24:15 crc kubenswrapper[4960]: I1002 09:24:15.382265 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zz9hm"] Oct 02 09:24:17 crc kubenswrapper[4960]: I1002 09:24:17.288892 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zz9hm" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="registry-server" containerID="cri-o://ea376c0dfa01457872c0504ae352c3cd4f665564ea4e0f440135c830de648c5a" gracePeriod=2 Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.302366 4960 generic.go:334] "Generic (PLEG): container finished" podID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerID="ea376c0dfa01457872c0504ae352c3cd4f665564ea4e0f440135c830de648c5a" exitCode=0 Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.302532 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zz9hm" event={"ID":"4ea495fa-92fb-463e-a12e-2a134f84b0ea","Type":"ContainerDied","Data":"ea376c0dfa01457872c0504ae352c3cd4f665564ea4e0f440135c830de648c5a"} Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.303148 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zz9hm" event={"ID":"4ea495fa-92fb-463e-a12e-2a134f84b0ea","Type":"ContainerDied","Data":"d1a2005d9b19bcf0edec697a96bdd7f058fb7d25e01582d92afe7b8a150df252"} Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.303169 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a2005d9b19bcf0edec697a96bdd7f058fb7d25e01582d92afe7b8a150df252" Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.328753 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.438045 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-catalog-content\") pod \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.438242 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kff4v\" (UniqueName: \"kubernetes.io/projected/4ea495fa-92fb-463e-a12e-2a134f84b0ea-kube-api-access-kff4v\") pod \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.438310 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-utilities\") pod \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\" (UID: \"4ea495fa-92fb-463e-a12e-2a134f84b0ea\") " Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.439604 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-utilities" (OuterVolumeSpecName: "utilities") pod "4ea495fa-92fb-463e-a12e-2a134f84b0ea" (UID: "4ea495fa-92fb-463e-a12e-2a134f84b0ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.446766 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ea495fa-92fb-463e-a12e-2a134f84b0ea-kube-api-access-kff4v" (OuterVolumeSpecName: "kube-api-access-kff4v") pod "4ea495fa-92fb-463e-a12e-2a134f84b0ea" (UID: "4ea495fa-92fb-463e-a12e-2a134f84b0ea"). InnerVolumeSpecName "kube-api-access-kff4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.489670 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ea495fa-92fb-463e-a12e-2a134f84b0ea" (UID: "4ea495fa-92fb-463e-a12e-2a134f84b0ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.542236 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.542288 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kff4v\" (UniqueName: \"kubernetes.io/projected/4ea495fa-92fb-463e-a12e-2a134f84b0ea-kube-api-access-kff4v\") on node \"crc\" DevicePath \"\"" Oct 02 09:24:18 crc kubenswrapper[4960]: I1002 09:24:18.542301 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ea495fa-92fb-463e-a12e-2a134f84b0ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:24:19 crc kubenswrapper[4960]: I1002 09:24:19.312792 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zz9hm" Oct 02 09:24:19 crc kubenswrapper[4960]: I1002 09:24:19.355787 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zz9hm"] Oct 02 09:24:19 crc kubenswrapper[4960]: I1002 09:24:19.365072 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zz9hm"] Oct 02 09:24:20 crc kubenswrapper[4960]: I1002 09:24:20.346713 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" path="/var/lib/kubelet/pods/4ea495fa-92fb-463e-a12e-2a134f84b0ea/volumes" Oct 02 09:24:59 crc kubenswrapper[4960]: I1002 09:24:59.150591 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:24:59 crc kubenswrapper[4960]: I1002 09:24:59.151424 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:25:27 crc kubenswrapper[4960]: I1002 09:25:27.011135 4960 generic.go:334] "Generic (PLEG): container finished" podID="aed3697d-fab0-4c3e-b5d0-21fb0037a412" containerID="eedcd7a8b6bba3120d5e7cc924e5e82f258f2541e8e637a43675452ca77dc609" exitCode=0 Oct 02 09:25:27 crc kubenswrapper[4960]: I1002 09:25:27.011235 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"aed3697d-fab0-4c3e-b5d0-21fb0037a412","Type":"ContainerDied","Data":"eedcd7a8b6bba3120d5e7cc924e5e82f258f2541e8e637a43675452ca77dc609"} Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.664264 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.708656 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ssh-key\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.708756 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ca-certs\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.708815 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ceph\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.708861 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-temporary\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.708906 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-workdir\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.708934 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config-secret\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.709045 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-config-data\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.709084 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.709124 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.709157 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-428rb\" (UniqueName: \"kubernetes.io/projected/aed3697d-fab0-4c3e-b5d0-21fb0037a412-kube-api-access-428rb\") pod \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\" (UID: \"aed3697d-fab0-4c3e-b5d0-21fb0037a412\") " Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.713780 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-config-data" (OuterVolumeSpecName: "config-data") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.714472 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.724174 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.724375 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed3697d-fab0-4c3e-b5d0-21fb0037a412-kube-api-access-428rb" (OuterVolumeSpecName: "kube-api-access-428rb") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "kube-api-access-428rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.729899 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.732812 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ceph" (OuterVolumeSpecName: "ceph") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.743698 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.745404 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.745663 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: E1002 09:25:28.746006 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="extract-utilities" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.746028 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="extract-utilities" Oct 02 09:25:28 crc kubenswrapper[4960]: E1002 09:25:28.746050 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="registry-server" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.746083 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="registry-server" Oct 02 09:25:28 crc kubenswrapper[4960]: E1002 09:25:28.746111 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="extract-content" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.746120 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="extract-content" Oct 02 09:25:28 crc kubenswrapper[4960]: E1002 09:25:28.746181 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed3697d-fab0-4c3e-b5d0-21fb0037a412" containerName="tempest-tests-tempest-tests-runner" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.746191 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed3697d-fab0-4c3e-b5d0-21fb0037a412" containerName="tempest-tests-tempest-tests-runner" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.746523 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ea495fa-92fb-463e-a12e-2a134f84b0ea" containerName="registry-server" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.746592 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed3697d-fab0-4c3e-b5d0-21fb0037a412" containerName="tempest-tests-tempest-tests-runner" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.747853 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.750909 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.751751 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.759806 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.775338 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.796301 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "aed3697d-fab0-4c3e-b5d0-21fb0037a412" (UID: "aed3697d-fab0-4c3e-b5d0-21fb0037a412"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.811804 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.811870 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.811920 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812028 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812082 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812132 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812196 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86dcw\" (UniqueName: \"kubernetes.io/projected/3cf44a73-42f7-41cc-82e4-f085f154cdbd-kube-api-access-86dcw\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812241 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812326 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812652 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812838 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812918 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.812962 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.813010 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/aed3697d-fab0-4c3e-b5d0-21fb0037a412-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.813025 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.813048 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aed3697d-fab0-4c3e-b5d0-21fb0037a412-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.813063 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-428rb\" (UniqueName: \"kubernetes.io/projected/aed3697d-fab0-4c3e-b5d0-21fb0037a412-kube-api-access-428rb\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.813074 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.813083 4960 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/aed3697d-fab0-4c3e-b5d0-21fb0037a412-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.860414 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.913939 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914014 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914049 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914094 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86dcw\" (UniqueName: \"kubernetes.io/projected/3cf44a73-42f7-41cc-82e4-f085f154cdbd-kube-api-access-86dcw\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914142 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914183 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914218 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914242 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914270 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.914637 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.915428 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.916164 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.919401 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.919480 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.919555 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.920509 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.923231 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:28 crc kubenswrapper[4960]: I1002 09:25:28.932597 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86dcw\" (UniqueName: \"kubernetes.io/projected/3cf44a73-42f7-41cc-82e4-f085f154cdbd-kube-api-access-86dcw\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:29 crc kubenswrapper[4960]: I1002 09:25:29.038771 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"aed3697d-fab0-4c3e-b5d0-21fb0037a412","Type":"ContainerDied","Data":"e6ed1391854bf9612376cd0f4ac9102c01b00222994ebdfee93b4f0d6788212c"} Oct 02 09:25:29 crc kubenswrapper[4960]: I1002 09:25:29.039455 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6ed1391854bf9612376cd0f4ac9102c01b00222994ebdfee93b4f0d6788212c" Oct 02 09:25:29 crc kubenswrapper[4960]: I1002 09:25:29.039595 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 02 09:25:29 crc kubenswrapper[4960]: I1002 09:25:29.150317 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:25:29 crc kubenswrapper[4960]: I1002 09:25:29.150413 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:25:29 crc kubenswrapper[4960]: I1002 09:25:29.171262 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:25:29 crc kubenswrapper[4960]: I1002 09:25:29.738603 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 02 09:25:30 crc kubenswrapper[4960]: I1002 09:25:30.054589 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"3cf44a73-42f7-41cc-82e4-f085f154cdbd","Type":"ContainerStarted","Data":"1055a916dff44782efd97d664b6d84280f0a4ff922130b904855986f301ad42f"} Oct 02 09:25:31 crc kubenswrapper[4960]: I1002 09:25:31.072127 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"3cf44a73-42f7-41cc-82e4-f085f154cdbd","Type":"ContainerStarted","Data":"f50c23cc7cbad040fb62c91f52fd4c79bddfb39b55097060006327fe852510c0"} Oct 02 09:25:31 crc kubenswrapper[4960]: I1002 09:25:31.095532 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-test" podStartSLOduration=3.095503173 podStartE2EDuration="3.095503173s" podCreationTimestamp="2025-10-02 09:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:25:31.092929444 +0000 UTC m=+7752.124875731" watchObservedRunningTime="2025-10-02 09:25:31.095503173 +0000 UTC m=+7752.127449460" Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.150021 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.151035 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.151121 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.152632 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.152720 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" gracePeriod=600 Oct 02 09:25:59 crc kubenswrapper[4960]: E1002 09:25:59.300490 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.381476 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" exitCode=0 Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.381556 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7"} Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.381641 4960 scope.go:117] "RemoveContainer" containerID="e25a36f3fbf7aec29094839309be354cbf21ad21e0ffe21d97b0d57f27fae0b5" Oct 02 09:25:59 crc kubenswrapper[4960]: I1002 09:25:59.382651 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:25:59 crc kubenswrapper[4960]: E1002 09:25:59.383209 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:26:10 crc kubenswrapper[4960]: I1002 09:26:10.337904 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:26:10 crc kubenswrapper[4960]: E1002 09:26:10.339509 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:26:23 crc kubenswrapper[4960]: I1002 09:26:23.330116 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:26:23 crc kubenswrapper[4960]: E1002 09:26:23.330898 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:26:36 crc kubenswrapper[4960]: I1002 09:26:36.330983 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:26:36 crc kubenswrapper[4960]: E1002 09:26:36.332093 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:26:51 crc kubenswrapper[4960]: I1002 09:26:51.330377 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:26:51 crc kubenswrapper[4960]: E1002 09:26:51.331387 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:27:02 crc kubenswrapper[4960]: I1002 09:27:02.333965 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:27:02 crc kubenswrapper[4960]: E1002 09:27:02.335003 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:27:13 crc kubenswrapper[4960]: I1002 09:27:13.329583 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:27:13 crc kubenswrapper[4960]: E1002 09:27:13.331943 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:27:24 crc kubenswrapper[4960]: I1002 09:27:24.329963 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:27:24 crc kubenswrapper[4960]: E1002 09:27:24.331133 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:27:37 crc kubenswrapper[4960]: I1002 09:27:37.331521 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:27:37 crc kubenswrapper[4960]: E1002 09:27:37.332839 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:27:51 crc kubenswrapper[4960]: I1002 09:27:51.330479 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:27:51 crc kubenswrapper[4960]: E1002 09:27:51.331676 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:28:05 crc kubenswrapper[4960]: I1002 09:28:05.330822 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:28:05 crc kubenswrapper[4960]: E1002 09:28:05.331904 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:28:17 crc kubenswrapper[4960]: I1002 09:28:17.331510 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:28:17 crc kubenswrapper[4960]: E1002 09:28:17.332996 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:28:28 crc kubenswrapper[4960]: I1002 09:28:28.331941 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:28:28 crc kubenswrapper[4960]: E1002 09:28:28.333061 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:28:42 crc kubenswrapper[4960]: I1002 09:28:42.330775 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:28:42 crc kubenswrapper[4960]: E1002 09:28:42.331916 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:28:55 crc kubenswrapper[4960]: I1002 09:28:55.330796 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:28:55 crc kubenswrapper[4960]: E1002 09:28:55.331812 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:29:09 crc kubenswrapper[4960]: I1002 09:29:09.330456 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:29:09 crc kubenswrapper[4960]: E1002 09:29:09.332247 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:29:20 crc kubenswrapper[4960]: I1002 09:29:20.337746 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:29:20 crc kubenswrapper[4960]: E1002 09:29:20.338724 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:29:35 crc kubenswrapper[4960]: I1002 09:29:35.330958 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:29:35 crc kubenswrapper[4960]: E1002 09:29:35.332146 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:29:50 crc kubenswrapper[4960]: I1002 09:29:50.342650 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:29:50 crc kubenswrapper[4960]: E1002 09:29:50.343732 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.158294 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2"] Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.161316 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.165548 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.165744 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.173177 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2"] Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.319231 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ztj7\" (UniqueName: \"kubernetes.io/projected/a1ec9501-885c-4e55-b572-eabb638fe147-kube-api-access-8ztj7\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.319342 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1ec9501-885c-4e55-b572-eabb638fe147-secret-volume\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.319393 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1ec9501-885c-4e55-b572-eabb638fe147-config-volume\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.421657 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ztj7\" (UniqueName: \"kubernetes.io/projected/a1ec9501-885c-4e55-b572-eabb638fe147-kube-api-access-8ztj7\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.421763 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1ec9501-885c-4e55-b572-eabb638fe147-secret-volume\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.421830 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1ec9501-885c-4e55-b572-eabb638fe147-config-volume\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.423282 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1ec9501-885c-4e55-b572-eabb638fe147-config-volume\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.433002 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1ec9501-885c-4e55-b572-eabb638fe147-secret-volume\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.449846 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ztj7\" (UniqueName: \"kubernetes.io/projected/a1ec9501-885c-4e55-b572-eabb638fe147-kube-api-access-8ztj7\") pod \"collect-profiles-29323290-r2mj2\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.506239 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:00 crc kubenswrapper[4960]: I1002 09:30:00.988359 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2"] Oct 02 09:30:01 crc kubenswrapper[4960]: I1002 09:30:01.942054 4960 generic.go:334] "Generic (PLEG): container finished" podID="a1ec9501-885c-4e55-b572-eabb638fe147" containerID="fd921851910738f31a1f3d3ce18ea2a9cc1b88cec694664c18cca9076cf25f95" exitCode=0 Oct 02 09:30:01 crc kubenswrapper[4960]: I1002 09:30:01.942157 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" event={"ID":"a1ec9501-885c-4e55-b572-eabb638fe147","Type":"ContainerDied","Data":"fd921851910738f31a1f3d3ce18ea2a9cc1b88cec694664c18cca9076cf25f95"} Oct 02 09:30:01 crc kubenswrapper[4960]: I1002 09:30:01.942683 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" event={"ID":"a1ec9501-885c-4e55-b572-eabb638fe147","Type":"ContainerStarted","Data":"b8e7c27c07df96471eeae626a51f6926bcc2d8ef60fd4a3478d9c146d59716df"} Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.349947 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.496569 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1ec9501-885c-4e55-b572-eabb638fe147-secret-volume\") pod \"a1ec9501-885c-4e55-b572-eabb638fe147\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.497288 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ztj7\" (UniqueName: \"kubernetes.io/projected/a1ec9501-885c-4e55-b572-eabb638fe147-kube-api-access-8ztj7\") pod \"a1ec9501-885c-4e55-b572-eabb638fe147\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.497406 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1ec9501-885c-4e55-b572-eabb638fe147-config-volume\") pod \"a1ec9501-885c-4e55-b572-eabb638fe147\" (UID: \"a1ec9501-885c-4e55-b572-eabb638fe147\") " Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.498300 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ec9501-885c-4e55-b572-eabb638fe147-config-volume" (OuterVolumeSpecName: "config-volume") pod "a1ec9501-885c-4e55-b572-eabb638fe147" (UID: "a1ec9501-885c-4e55-b572-eabb638fe147"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.507595 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ec9501-885c-4e55-b572-eabb638fe147-kube-api-access-8ztj7" (OuterVolumeSpecName: "kube-api-access-8ztj7") pod "a1ec9501-885c-4e55-b572-eabb638fe147" (UID: "a1ec9501-885c-4e55-b572-eabb638fe147"). InnerVolumeSpecName "kube-api-access-8ztj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.510208 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ec9501-885c-4e55-b572-eabb638fe147-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a1ec9501-885c-4e55-b572-eabb638fe147" (UID: "a1ec9501-885c-4e55-b572-eabb638fe147"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.600120 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1ec9501-885c-4e55-b572-eabb638fe147-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.600210 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1ec9501-885c-4e55-b572-eabb638fe147-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.600227 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ztj7\" (UniqueName: \"kubernetes.io/projected/a1ec9501-885c-4e55-b572-eabb638fe147-kube-api-access-8ztj7\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.969291 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" event={"ID":"a1ec9501-885c-4e55-b572-eabb638fe147","Type":"ContainerDied","Data":"b8e7c27c07df96471eeae626a51f6926bcc2d8ef60fd4a3478d9c146d59716df"} Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.969340 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8e7c27c07df96471eeae626a51f6926bcc2d8ef60fd4a3478d9c146d59716df" Oct 02 09:30:03 crc kubenswrapper[4960]: I1002 09:30:03.969379 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-r2mj2" Oct 02 09:30:04 crc kubenswrapper[4960]: I1002 09:30:04.330615 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:30:04 crc kubenswrapper[4960]: E1002 09:30:04.331215 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:30:04 crc kubenswrapper[4960]: I1002 09:30:04.435160 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x"] Oct 02 09:30:04 crc kubenswrapper[4960]: I1002 09:30:04.444504 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-5495x"] Oct 02 09:30:04 crc kubenswrapper[4960]: I1002 09:30:04.766062 4960 scope.go:117] "RemoveContainer" containerID="02d243db984973a4108e203c544f86c856842f0f6d0a40b65fde6b347ccb444a" Oct 02 09:30:04 crc kubenswrapper[4960]: I1002 09:30:04.793387 4960 scope.go:117] "RemoveContainer" containerID="bd764d472f75fdab75af95ca54ad6100afdafd5a83cd48a809b50b86a17bccee" Oct 02 09:30:06 crc kubenswrapper[4960]: I1002 09:30:06.346564 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64343f3-cd2f-4722-a9d9-2d12360dc6f5" path="/var/lib/kubelet/pods/d64343f3-cd2f-4722-a9d9-2d12360dc6f5/volumes" Oct 02 09:30:16 crc kubenswrapper[4960]: I1002 09:30:16.329885 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:30:16 crc kubenswrapper[4960]: E1002 09:30:16.330887 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:30:23 crc kubenswrapper[4960]: I1002 09:30:23.157708 4960 generic.go:334] "Generic (PLEG): container finished" podID="3cf44a73-42f7-41cc-82e4-f085f154cdbd" containerID="f50c23cc7cbad040fb62c91f52fd4c79bddfb39b55097060006327fe852510c0" exitCode=0 Oct 02 09:30:23 crc kubenswrapper[4960]: I1002 09:30:23.157803 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"3cf44a73-42f7-41cc-82e4-f085f154cdbd","Type":"ContainerDied","Data":"f50c23cc7cbad040fb62c91f52fd4c79bddfb39b55097060006327fe852510c0"} Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.754677 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918239 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-temporary\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918427 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-config-data\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918580 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86dcw\" (UniqueName: \"kubernetes.io/projected/3cf44a73-42f7-41cc-82e4-f085f154cdbd-kube-api-access-86dcw\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918620 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config-secret\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918648 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ssh-key\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918673 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ceph\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918817 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ca-certs\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918856 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-workdir\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918900 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.918993 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.919067 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config\") pod \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\" (UID: \"3cf44a73-42f7-41cc-82e4-f085f154cdbd\") " Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.919320 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-config-data" (OuterVolumeSpecName: "config-data") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.919642 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.919662 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.925067 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ceph" (OuterVolumeSpecName: "ceph") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.925602 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf44a73-42f7-41cc-82e4-f085f154cdbd-kube-api-access-86dcw" (OuterVolumeSpecName: "kube-api-access-86dcw") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "kube-api-access-86dcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.926950 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.929520 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.967291 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.974255 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.998397 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:30:24 crc kubenswrapper[4960]: I1002 09:30:24.999745 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3cf44a73-42f7-41cc-82e4-f085f154cdbd" (UID: "3cf44a73-42f7-41cc-82e4-f085f154cdbd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022267 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86dcw\" (UniqueName: \"kubernetes.io/projected/3cf44a73-42f7-41cc-82e4-f085f154cdbd-kube-api-access-86dcw\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022309 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022329 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022341 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022352 4960 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3cf44a73-42f7-41cc-82e4-f085f154cdbd-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022365 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3cf44a73-42f7-41cc-82e4-f085f154cdbd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022415 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.022431 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3cf44a73-42f7-41cc-82e4-f085f154cdbd-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.043861 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.123928 4960 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.179886 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"3cf44a73-42f7-41cc-82e4-f085f154cdbd","Type":"ContainerDied","Data":"1055a916dff44782efd97d664b6d84280f0a4ff922130b904855986f301ad42f"} Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.180261 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1055a916dff44782efd97d664b6d84280f0a4ff922130b904855986f301ad42f" Oct 02 09:30:25 crc kubenswrapper[4960]: I1002 09:30:25.180046 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 02 09:30:31 crc kubenswrapper[4960]: I1002 09:30:31.331656 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:30:31 crc kubenswrapper[4960]: E1002 09:30:31.333055 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.054364 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 09:30:32 crc kubenswrapper[4960]: E1002 09:30:32.054942 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf44a73-42f7-41cc-82e4-f085f154cdbd" containerName="tempest-tests-tempest-tests-runner" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.054995 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf44a73-42f7-41cc-82e4-f085f154cdbd" containerName="tempest-tests-tempest-tests-runner" Oct 02 09:30:32 crc kubenswrapper[4960]: E1002 09:30:32.055071 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ec9501-885c-4e55-b572-eabb638fe147" containerName="collect-profiles" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.055091 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ec9501-885c-4e55-b572-eabb638fe147" containerName="collect-profiles" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.055379 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ec9501-885c-4e55-b572-eabb638fe147" containerName="collect-profiles" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.055406 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf44a73-42f7-41cc-82e4-f085f154cdbd" containerName="tempest-tests-tempest-tests-runner" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.056400 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.060768 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vk2jb" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.073202 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.170945 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np7ht\" (UniqueName: \"kubernetes.io/projected/06a6e149-1b05-45a2-98b0-f53b766d60eb-kube-api-access-np7ht\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"06a6e149-1b05-45a2-98b0-f53b766d60eb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.171473 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"06a6e149-1b05-45a2-98b0-f53b766d60eb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.274130 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"06a6e149-1b05-45a2-98b0-f53b766d60eb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.274251 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np7ht\" (UniqueName: \"kubernetes.io/projected/06a6e149-1b05-45a2-98b0-f53b766d60eb-kube-api-access-np7ht\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"06a6e149-1b05-45a2-98b0-f53b766d60eb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.275085 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"06a6e149-1b05-45a2-98b0-f53b766d60eb\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.294134 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np7ht\" (UniqueName: \"kubernetes.io/projected/06a6e149-1b05-45a2-98b0-f53b766d60eb-kube-api-access-np7ht\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"06a6e149-1b05-45a2-98b0-f53b766d60eb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.304512 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"06a6e149-1b05-45a2-98b0-f53b766d60eb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.397134 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.876524 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:30:32 crc kubenswrapper[4960]: I1002 09:30:32.877402 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 09:30:33 crc kubenswrapper[4960]: I1002 09:30:33.277393 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"06a6e149-1b05-45a2-98b0-f53b766d60eb","Type":"ContainerStarted","Data":"938012df4584f7b5977ad690ba2735357e82a2ebe1b290c41baf98277d35752c"} Oct 02 09:30:34 crc kubenswrapper[4960]: I1002 09:30:34.289521 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"06a6e149-1b05-45a2-98b0-f53b766d60eb","Type":"ContainerStarted","Data":"d0ffad936f0d8dfa78e31266cd70de82343aaa2aecdcdeac384177876c3bbad9"} Oct 02 09:30:34 crc kubenswrapper[4960]: I1002 09:30:34.309316 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.326214856 podStartE2EDuration="2.309293525s" podCreationTimestamp="2025-10-02 09:30:32 +0000 UTC" firstStartedPulling="2025-10-02 09:30:32.876317684 +0000 UTC m=+8053.908263971" lastFinishedPulling="2025-10-02 09:30:33.859396353 +0000 UTC m=+8054.891342640" observedRunningTime="2025-10-02 09:30:34.307117147 +0000 UTC m=+8055.339063444" watchObservedRunningTime="2025-10-02 09:30:34.309293525 +0000 UTC m=+8055.341239802" Oct 02 09:30:43 crc kubenswrapper[4960]: I1002 09:30:43.329843 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:30:43 crc kubenswrapper[4960]: E1002 09:30:43.330727 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.672548 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.675096 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.678421 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.678439 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-config" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.678605 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-private-key" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.678834 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"tobiko-secret" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.679064 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-public-key" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.723156 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.812921 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813004 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813039 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813077 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813099 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813118 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813162 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813201 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813231 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813262 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813373 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.813964 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql2kx\" (UniqueName: \"kubernetes.io/projected/2b480863-210b-4fb7-8980-e939d5fc9603-kube-api-access-ql2kx\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.916592 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.916669 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.916698 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.916756 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.916811 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.916857 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.916898 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.917207 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.917660 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql2kx\" (UniqueName: \"kubernetes.io/projected/2b480863-210b-4fb7-8980-e939d5fc9603-kube-api-access-ql2kx\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.918055 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.918135 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.918230 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.919381 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.919387 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.919737 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.919924 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.919953 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.919928 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.921886 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.926050 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.926921 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.928485 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.928529 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.939425 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql2kx\" (UniqueName: \"kubernetes.io/projected/2b480863-210b-4fb7-8980-e939d5fc9603-kube-api-access-ql2kx\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:55 crc kubenswrapper[4960]: I1002 09:30:55.954135 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:56 crc kubenswrapper[4960]: I1002 09:30:56.015741 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:30:56 crc kubenswrapper[4960]: I1002 09:30:56.330994 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:30:56 crc kubenswrapper[4960]: E1002 09:30:56.331566 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:30:56 crc kubenswrapper[4960]: I1002 09:30:56.627517 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 02 09:30:57 crc kubenswrapper[4960]: I1002 09:30:57.559792 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"2b480863-210b-4fb7-8980-e939d5fc9603","Type":"ContainerStarted","Data":"e4c725a82405e230b8ccc0fe2ad4afbded74dfdaba5366b54ea32b672ac5fe06"} Oct 02 09:31:04 crc kubenswrapper[4960]: I1002 09:31:04.897380 4960 scope.go:117] "RemoveContainer" containerID="20962b7aff238872d3935749a500fcf85ace57b25018940be68c9777345c0000" Oct 02 09:31:04 crc kubenswrapper[4960]: I1002 09:31:04.926328 4960 scope.go:117] "RemoveContainer" containerID="ea376c0dfa01457872c0504ae352c3cd4f665564ea4e0f440135c830de648c5a" Oct 02 09:31:09 crc kubenswrapper[4960]: I1002 09:31:09.331783 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:31:09 crc kubenswrapper[4960]: I1002 09:31:09.700315 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"3e3b98926708e1815f9439704caba866fca62e36285e3709b97655ae40054844"} Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.468911 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-46rjf"] Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.472178 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.481536 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-46rjf"] Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.545421 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-utilities\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.545480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rbd8\" (UniqueName: \"kubernetes.io/projected/7c688f41-d7c8-4c7e-abde-803036bb7423-kube-api-access-8rbd8\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.545600 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-catalog-content\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.646723 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-utilities\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.646791 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rbd8\" (UniqueName: \"kubernetes.io/projected/7c688f41-d7c8-4c7e-abde-803036bb7423-kube-api-access-8rbd8\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.646868 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-catalog-content\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.647275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-utilities\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.647298 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-catalog-content\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.670912 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rbd8\" (UniqueName: \"kubernetes.io/projected/7c688f41-d7c8-4c7e-abde-803036bb7423-kube-api-access-8rbd8\") pod \"redhat-operators-46rjf\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:29 crc kubenswrapper[4960]: I1002 09:31:29.798405 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:30 crc kubenswrapper[4960]: I1002 09:31:30.143396 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-46rjf"] Oct 02 09:31:42 crc kubenswrapper[4960]: E1002 09:31:42.644569 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tobiko:current-podified" Oct 02 09:31:42 crc kubenswrapper[4960]: E1002 09:31:42.645356 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tobiko-tests-tobiko,Image:quay.io/podified-antelope-centos9/openstack-tobiko:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TOBIKO_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:TOBIKO_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:TOBIKO_LOGS_DIR_NAME,Value:tobiko-tests-tobiko-s00-podified-functional,ValueFrom:nil,},EnvVar{Name:TOBIKO_PYTEST_ADDOPTS,Value:,ValueFrom:nil,},EnvVar{Name:TOBIKO_TESTENV,Value:functional -- tobiko/tests/functional/podified/test_topology.py,ValueFrom:nil,},EnvVar{Name:TOBIKO_VERSION,Value:master,ValueFrom:nil,},EnvVar{Name:TOX_NUM_PROCESSES,Value:2,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{8 0} {} 8 DecimalSI},memory: {{8589934592 0} {} BinarySI},},Requests:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tobiko,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tobiko/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/tobiko/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-config,ReadOnly:false,MountPath:/etc/tobiko/tobiko.conf,SubPath:tobiko.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-private-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa,SubPath:id_ecdsa,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-public-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa.pub,SubPath:id_ecdsa.pub,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kubeconfig,ReadOnly:true,MountPath:/var/lib/tobiko/.kube/config,SubPath:config,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ql2kx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42495,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42495,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tobiko-tests-tobiko-s00-podified-functional_openstack(2b480863-210b-4fb7-8980-e939d5fc9603): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:31:42 crc kubenswrapper[4960]: E1002 09:31:42.646657 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="2b480863-210b-4fb7-8980-e939d5fc9603" Oct 02 09:31:43 crc kubenswrapper[4960]: I1002 09:31:43.032235 4960 generic.go:334] "Generic (PLEG): container finished" podID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerID="9149c94af2fd78ca4e3bf99df5b70faa303e2222200fec19a638bf503caf06b1" exitCode=0 Oct 02 09:31:43 crc kubenswrapper[4960]: I1002 09:31:43.032318 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46rjf" event={"ID":"7c688f41-d7c8-4c7e-abde-803036bb7423","Type":"ContainerDied","Data":"9149c94af2fd78ca4e3bf99df5b70faa303e2222200fec19a638bf503caf06b1"} Oct 02 09:31:43 crc kubenswrapper[4960]: I1002 09:31:43.032405 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46rjf" event={"ID":"7c688f41-d7c8-4c7e-abde-803036bb7423","Type":"ContainerStarted","Data":"7c0487432228e58605ded2d76b47cacf099ee4f08dc585f5c74780c3ddd433de"} Oct 02 09:31:43 crc kubenswrapper[4960]: E1002 09:31:43.035193 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tobiko:current-podified\\\"\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="2b480863-210b-4fb7-8980-e939d5fc9603" Oct 02 09:31:45 crc kubenswrapper[4960]: I1002 09:31:45.053512 4960 generic.go:334] "Generic (PLEG): container finished" podID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerID="a4a55df828cb3e044d16a2fc8dca1e8dc11761257ac000a0b28439b52b53db81" exitCode=0 Oct 02 09:31:45 crc kubenswrapper[4960]: I1002 09:31:45.053628 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46rjf" event={"ID":"7c688f41-d7c8-4c7e-abde-803036bb7423","Type":"ContainerDied","Data":"a4a55df828cb3e044d16a2fc8dca1e8dc11761257ac000a0b28439b52b53db81"} Oct 02 09:31:47 crc kubenswrapper[4960]: I1002 09:31:47.077781 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46rjf" event={"ID":"7c688f41-d7c8-4c7e-abde-803036bb7423","Type":"ContainerStarted","Data":"32893daf8b5808a5dbf6af4f7d1890ecc698202ff358862126177d5dbd0e3c0f"} Oct 02 09:31:47 crc kubenswrapper[4960]: I1002 09:31:47.108647 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-46rjf" podStartSLOduration=15.034047943 podStartE2EDuration="18.108621551s" podCreationTimestamp="2025-10-02 09:31:29 +0000 UTC" firstStartedPulling="2025-10-02 09:31:43.035293518 +0000 UTC m=+8124.067239805" lastFinishedPulling="2025-10-02 09:31:46.109867126 +0000 UTC m=+8127.141813413" observedRunningTime="2025-10-02 09:31:47.103040583 +0000 UTC m=+8128.134986880" watchObservedRunningTime="2025-10-02 09:31:47.108621551 +0000 UTC m=+8128.140567838" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.428209 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lgfr6"] Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.431427 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.441201 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lgfr6"] Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.539002 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95bkm\" (UniqueName: \"kubernetes.io/projected/9086d1db-befd-41b9-831f-7782055fb295-kube-api-access-95bkm\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.539075 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-utilities\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.539214 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-catalog-content\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.642040 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-catalog-content\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.642223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95bkm\" (UniqueName: \"kubernetes.io/projected/9086d1db-befd-41b9-831f-7782055fb295-kube-api-access-95bkm\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.642275 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-utilities\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.642707 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-catalog-content\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.642821 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-utilities\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.671125 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95bkm\" (UniqueName: \"kubernetes.io/projected/9086d1db-befd-41b9-831f-7782055fb295-kube-api-access-95bkm\") pod \"community-operators-lgfr6\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:48 crc kubenswrapper[4960]: I1002 09:31:48.765586 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:49 crc kubenswrapper[4960]: I1002 09:31:49.367811 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lgfr6"] Oct 02 09:31:49 crc kubenswrapper[4960]: W1002 09:31:49.367937 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9086d1db_befd_41b9_831f_7782055fb295.slice/crio-e4bc66470d8e56a2982c98e7a1bb7c0043d0123a7bfacf7a97206ebfc98850bf WatchSource:0}: Error finding container e4bc66470d8e56a2982c98e7a1bb7c0043d0123a7bfacf7a97206ebfc98850bf: Status 404 returned error can't find the container with id e4bc66470d8e56a2982c98e7a1bb7c0043d0123a7bfacf7a97206ebfc98850bf Oct 02 09:31:49 crc kubenswrapper[4960]: I1002 09:31:49.798560 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:49 crc kubenswrapper[4960]: I1002 09:31:49.798813 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:50 crc kubenswrapper[4960]: I1002 09:31:50.111249 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerStarted","Data":"7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475"} Oct 02 09:31:50 crc kubenswrapper[4960]: I1002 09:31:50.111306 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerStarted","Data":"e4bc66470d8e56a2982c98e7a1bb7c0043d0123a7bfacf7a97206ebfc98850bf"} Oct 02 09:31:50 crc kubenswrapper[4960]: I1002 09:31:50.848400 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-46rjf" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="registry-server" probeResult="failure" output=< Oct 02 09:31:50 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 09:31:50 crc kubenswrapper[4960]: > Oct 02 09:31:51 crc kubenswrapper[4960]: I1002 09:31:51.124820 4960 generic.go:334] "Generic (PLEG): container finished" podID="9086d1db-befd-41b9-831f-7782055fb295" containerID="7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475" exitCode=0 Oct 02 09:31:51 crc kubenswrapper[4960]: I1002 09:31:51.125199 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerDied","Data":"7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475"} Oct 02 09:31:52 crc kubenswrapper[4960]: I1002 09:31:52.137028 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerStarted","Data":"211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747"} Oct 02 09:31:53 crc kubenswrapper[4960]: I1002 09:31:53.153426 4960 generic.go:334] "Generic (PLEG): container finished" podID="9086d1db-befd-41b9-831f-7782055fb295" containerID="211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747" exitCode=0 Oct 02 09:31:53 crc kubenswrapper[4960]: I1002 09:31:53.153530 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerDied","Data":"211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747"} Oct 02 09:31:54 crc kubenswrapper[4960]: I1002 09:31:54.171083 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerStarted","Data":"cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea"} Oct 02 09:31:54 crc kubenswrapper[4960]: I1002 09:31:54.195413 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lgfr6" podStartSLOduration=3.741647637 podStartE2EDuration="6.195394399s" podCreationTimestamp="2025-10-02 09:31:48 +0000 UTC" firstStartedPulling="2025-10-02 09:31:51.127904278 +0000 UTC m=+8132.159850565" lastFinishedPulling="2025-10-02 09:31:53.58165105 +0000 UTC m=+8134.613597327" observedRunningTime="2025-10-02 09:31:54.193645832 +0000 UTC m=+8135.225592139" watchObservedRunningTime="2025-10-02 09:31:54.195394399 +0000 UTC m=+8135.227340686" Oct 02 09:31:58 crc kubenswrapper[4960]: I1002 09:31:58.766231 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:58 crc kubenswrapper[4960]: I1002 09:31:58.767196 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:58 crc kubenswrapper[4960]: I1002 09:31:58.821434 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:59 crc kubenswrapper[4960]: I1002 09:31:59.293260 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:31:59 crc kubenswrapper[4960]: I1002 09:31:59.348493 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lgfr6"] Oct 02 09:31:59 crc kubenswrapper[4960]: I1002 09:31:59.852620 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:31:59 crc kubenswrapper[4960]: I1002 09:31:59.913561 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:32:00 crc kubenswrapper[4960]: I1002 09:32:00.245794 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"2b480863-210b-4fb7-8980-e939d5fc9603","Type":"ContainerStarted","Data":"3977b4bb2a3ade5145c826bb4a2a80db8dbc3508c6ae0461ad8b3624127c6098"} Oct 02 09:32:00 crc kubenswrapper[4960]: I1002 09:32:00.281658 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podStartSLOduration=4.001238747 podStartE2EDuration="1m6.281622372s" podCreationTimestamp="2025-10-02 09:30:54 +0000 UTC" firstStartedPulling="2025-10-02 09:30:56.632138305 +0000 UTC m=+8077.664084582" lastFinishedPulling="2025-10-02 09:31:58.91252192 +0000 UTC m=+8139.944468207" observedRunningTime="2025-10-02 09:32:00.269347296 +0000 UTC m=+8141.301293583" watchObservedRunningTime="2025-10-02 09:32:00.281622372 +0000 UTC m=+8141.313568659" Oct 02 09:32:01 crc kubenswrapper[4960]: I1002 09:32:01.257900 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lgfr6" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="registry-server" containerID="cri-o://cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea" gracePeriod=2 Oct 02 09:32:01 crc kubenswrapper[4960]: I1002 09:32:01.471675 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-46rjf"] Oct 02 09:32:01 crc kubenswrapper[4960]: I1002 09:32:01.472142 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-46rjf" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="registry-server" containerID="cri-o://32893daf8b5808a5dbf6af4f7d1890ecc698202ff358862126177d5dbd0e3c0f" gracePeriod=2 Oct 02 09:32:02 crc kubenswrapper[4960]: I1002 09:32:02.895342 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.003304 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-utilities\") pod \"9086d1db-befd-41b9-831f-7782055fb295\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.003639 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95bkm\" (UniqueName: \"kubernetes.io/projected/9086d1db-befd-41b9-831f-7782055fb295-kube-api-access-95bkm\") pod \"9086d1db-befd-41b9-831f-7782055fb295\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.003738 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-catalog-content\") pod \"9086d1db-befd-41b9-831f-7782055fb295\" (UID: \"9086d1db-befd-41b9-831f-7782055fb295\") " Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.004598 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-utilities" (OuterVolumeSpecName: "utilities") pod "9086d1db-befd-41b9-831f-7782055fb295" (UID: "9086d1db-befd-41b9-831f-7782055fb295"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.012287 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9086d1db-befd-41b9-831f-7782055fb295-kube-api-access-95bkm" (OuterVolumeSpecName: "kube-api-access-95bkm") pod "9086d1db-befd-41b9-831f-7782055fb295" (UID: "9086d1db-befd-41b9-831f-7782055fb295"). InnerVolumeSpecName "kube-api-access-95bkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.082452 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9086d1db-befd-41b9-831f-7782055fb295" (UID: "9086d1db-befd-41b9-831f-7782055fb295"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.107575 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.107648 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9086d1db-befd-41b9-831f-7782055fb295-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.107671 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95bkm\" (UniqueName: \"kubernetes.io/projected/9086d1db-befd-41b9-831f-7782055fb295-kube-api-access-95bkm\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.288042 4960 generic.go:334] "Generic (PLEG): container finished" podID="9086d1db-befd-41b9-831f-7782055fb295" containerID="cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea" exitCode=0 Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.288507 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgfr6" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.292083 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerDied","Data":"cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea"} Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.292168 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgfr6" event={"ID":"9086d1db-befd-41b9-831f-7782055fb295","Type":"ContainerDied","Data":"e4bc66470d8e56a2982c98e7a1bb7c0043d0123a7bfacf7a97206ebfc98850bf"} Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.292196 4960 scope.go:117] "RemoveContainer" containerID="cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.301242 4960 generic.go:334] "Generic (PLEG): container finished" podID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerID="32893daf8b5808a5dbf6af4f7d1890ecc698202ff358862126177d5dbd0e3c0f" exitCode=0 Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.301287 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46rjf" event={"ID":"7c688f41-d7c8-4c7e-abde-803036bb7423","Type":"ContainerDied","Data":"32893daf8b5808a5dbf6af4f7d1890ecc698202ff358862126177d5dbd0e3c0f"} Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.319649 4960 scope.go:117] "RemoveContainer" containerID="211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.338088 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lgfr6"] Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.348047 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lgfr6"] Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.353580 4960 scope.go:117] "RemoveContainer" containerID="7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.449261 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.451882 4960 scope.go:117] "RemoveContainer" containerID="cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea" Oct 02 09:32:03 crc kubenswrapper[4960]: E1002 09:32:03.452181 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea\": container with ID starting with cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea not found: ID does not exist" containerID="cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.452221 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea"} err="failed to get container status \"cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea\": rpc error: code = NotFound desc = could not find container \"cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea\": container with ID starting with cdc4b0fb7262e14bc00e4bbb97107132783fd34012d3255d63074914ea866aea not found: ID does not exist" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.452244 4960 scope.go:117] "RemoveContainer" containerID="211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747" Oct 02 09:32:03 crc kubenswrapper[4960]: E1002 09:32:03.452466 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747\": container with ID starting with 211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747 not found: ID does not exist" containerID="211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.452581 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747"} err="failed to get container status \"211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747\": rpc error: code = NotFound desc = could not find container \"211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747\": container with ID starting with 211823c69c0048ce3bcd31b1b63b46343261ef6d3617d015ecf214ad2f76b747 not found: ID does not exist" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.452660 4960 scope.go:117] "RemoveContainer" containerID="7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475" Oct 02 09:32:03 crc kubenswrapper[4960]: E1002 09:32:03.453039 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475\": container with ID starting with 7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475 not found: ID does not exist" containerID="7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.453066 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475"} err="failed to get container status \"7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475\": rpc error: code = NotFound desc = could not find container \"7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475\": container with ID starting with 7da2c00ef528f93b5b129c8c8aaade4c0fbec460e3844ef770ee5bb73bfde475 not found: ID does not exist" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.518775 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-utilities\") pod \"7c688f41-d7c8-4c7e-abde-803036bb7423\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.519120 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rbd8\" (UniqueName: \"kubernetes.io/projected/7c688f41-d7c8-4c7e-abde-803036bb7423-kube-api-access-8rbd8\") pod \"7c688f41-d7c8-4c7e-abde-803036bb7423\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.519190 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-catalog-content\") pod \"7c688f41-d7c8-4c7e-abde-803036bb7423\" (UID: \"7c688f41-d7c8-4c7e-abde-803036bb7423\") " Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.520303 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-utilities" (OuterVolumeSpecName: "utilities") pod "7c688f41-d7c8-4c7e-abde-803036bb7423" (UID: "7c688f41-d7c8-4c7e-abde-803036bb7423"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.525754 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c688f41-d7c8-4c7e-abde-803036bb7423-kube-api-access-8rbd8" (OuterVolumeSpecName: "kube-api-access-8rbd8") pod "7c688f41-d7c8-4c7e-abde-803036bb7423" (UID: "7c688f41-d7c8-4c7e-abde-803036bb7423"). InnerVolumeSpecName "kube-api-access-8rbd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.613313 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c688f41-d7c8-4c7e-abde-803036bb7423" (UID: "7c688f41-d7c8-4c7e-abde-803036bb7423"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.622517 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.622559 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c688f41-d7c8-4c7e-abde-803036bb7423-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[4960]: I1002 09:32:03.622577 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rbd8\" (UniqueName: \"kubernetes.io/projected/7c688f41-d7c8-4c7e-abde-803036bb7423-kube-api-access-8rbd8\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.328989 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46rjf" Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.346575 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9086d1db-befd-41b9-831f-7782055fb295" path="/var/lib/kubelet/pods/9086d1db-befd-41b9-831f-7782055fb295/volumes" Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.348084 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46rjf" event={"ID":"7c688f41-d7c8-4c7e-abde-803036bb7423","Type":"ContainerDied","Data":"7c0487432228e58605ded2d76b47cacf099ee4f08dc585f5c74780c3ddd433de"} Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.351435 4960 scope.go:117] "RemoveContainer" containerID="32893daf8b5808a5dbf6af4f7d1890ecc698202ff358862126177d5dbd0e3c0f" Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.370316 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-46rjf"] Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.379542 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-46rjf"] Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.379931 4960 scope.go:117] "RemoveContainer" containerID="a4a55df828cb3e044d16a2fc8dca1e8dc11761257ac000a0b28439b52b53db81" Oct 02 09:32:04 crc kubenswrapper[4960]: I1002 09:32:04.407874 4960 scope.go:117] "RemoveContainer" containerID="9149c94af2fd78ca4e3bf99df5b70faa303e2222200fec19a638bf503caf06b1" Oct 02 09:32:06 crc kubenswrapper[4960]: I1002 09:32:06.345660 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" path="/var/lib/kubelet/pods/7c688f41-d7c8-4c7e-abde-803036bb7423/volumes" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.632886 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9fxdw"] Oct 02 09:32:08 crc kubenswrapper[4960]: E1002 09:32:08.633432 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="extract-utilities" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633454 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="extract-utilities" Oct 02 09:32:08 crc kubenswrapper[4960]: E1002 09:32:08.633475 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="extract-content" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633484 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="extract-content" Oct 02 09:32:08 crc kubenswrapper[4960]: E1002 09:32:08.633506 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="extract-utilities" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633517 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="extract-utilities" Oct 02 09:32:08 crc kubenswrapper[4960]: E1002 09:32:08.633543 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="registry-server" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633553 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="registry-server" Oct 02 09:32:08 crc kubenswrapper[4960]: E1002 09:32:08.633610 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="extract-content" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633618 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="extract-content" Oct 02 09:32:08 crc kubenswrapper[4960]: E1002 09:32:08.633640 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="registry-server" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633647 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="registry-server" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633908 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c688f41-d7c8-4c7e-abde-803036bb7423" containerName="registry-server" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.633930 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9086d1db-befd-41b9-831f-7782055fb295" containerName="registry-server" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.638447 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.655084 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fxdw"] Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.745196 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-utilities\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.745563 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fjzl\" (UniqueName: \"kubernetes.io/projected/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-kube-api-access-6fjzl\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.745749 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-catalog-content\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.847527 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-catalog-content\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.847610 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-utilities\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.847705 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fjzl\" (UniqueName: \"kubernetes.io/projected/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-kube-api-access-6fjzl\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.848346 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-catalog-content\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.848398 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-utilities\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.884054 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fjzl\" (UniqueName: \"kubernetes.io/projected/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-kube-api-access-6fjzl\") pod \"redhat-marketplace-9fxdw\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:08 crc kubenswrapper[4960]: I1002 09:32:08.963667 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:09 crc kubenswrapper[4960]: I1002 09:32:09.481746 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fxdw"] Oct 02 09:32:10 crc kubenswrapper[4960]: I1002 09:32:10.391438 4960 generic.go:334] "Generic (PLEG): container finished" podID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerID="4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf" exitCode=0 Oct 02 09:32:10 crc kubenswrapper[4960]: I1002 09:32:10.391535 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fxdw" event={"ID":"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0","Type":"ContainerDied","Data":"4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf"} Oct 02 09:32:10 crc kubenswrapper[4960]: I1002 09:32:10.395191 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fxdw" event={"ID":"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0","Type":"ContainerStarted","Data":"bf2812e705cc64e7dd393e6ff7e16e4053f437f85fe0cfd07abe0ac45b2928cc"} Oct 02 09:32:12 crc kubenswrapper[4960]: I1002 09:32:12.413573 4960 generic.go:334] "Generic (PLEG): container finished" podID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerID="2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7" exitCode=0 Oct 02 09:32:12 crc kubenswrapper[4960]: I1002 09:32:12.413702 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fxdw" event={"ID":"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0","Type":"ContainerDied","Data":"2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7"} Oct 02 09:32:13 crc kubenswrapper[4960]: I1002 09:32:13.429689 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fxdw" event={"ID":"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0","Type":"ContainerStarted","Data":"3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4"} Oct 02 09:32:13 crc kubenswrapper[4960]: I1002 09:32:13.459625 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9fxdw" podStartSLOduration=2.935313027 podStartE2EDuration="5.459594825s" podCreationTimestamp="2025-10-02 09:32:08 +0000 UTC" firstStartedPulling="2025-10-02 09:32:10.39420498 +0000 UTC m=+8151.426151267" lastFinishedPulling="2025-10-02 09:32:12.918486778 +0000 UTC m=+8153.950433065" observedRunningTime="2025-10-02 09:32:13.449367953 +0000 UTC m=+8154.481314230" watchObservedRunningTime="2025-10-02 09:32:13.459594825 +0000 UTC m=+8154.491541112" Oct 02 09:32:18 crc kubenswrapper[4960]: I1002 09:32:18.964087 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:18 crc kubenswrapper[4960]: I1002 09:32:18.964804 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:19 crc kubenswrapper[4960]: I1002 09:32:19.014647 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:19 crc kubenswrapper[4960]: I1002 09:32:19.536204 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:19 crc kubenswrapper[4960]: I1002 09:32:19.629593 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fxdw"] Oct 02 09:32:21 crc kubenswrapper[4960]: I1002 09:32:21.501392 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9fxdw" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="registry-server" containerID="cri-o://3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4" gracePeriod=2 Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.135783 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.297765 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fjzl\" (UniqueName: \"kubernetes.io/projected/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-kube-api-access-6fjzl\") pod \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.298280 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-catalog-content\") pod \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.298440 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-utilities\") pod \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\" (UID: \"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0\") " Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.300370 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-utilities" (OuterVolumeSpecName: "utilities") pod "a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" (UID: "a1c49a74-f205-4feb-b0ed-4dd32c4e99e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.305391 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-kube-api-access-6fjzl" (OuterVolumeSpecName: "kube-api-access-6fjzl") pod "a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" (UID: "a1c49a74-f205-4feb-b0ed-4dd32c4e99e0"). InnerVolumeSpecName "kube-api-access-6fjzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.313920 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" (UID: "a1c49a74-f205-4feb-b0ed-4dd32c4e99e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.400880 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.400922 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fjzl\" (UniqueName: \"kubernetes.io/projected/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-kube-api-access-6fjzl\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.400935 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.516237 4960 generic.go:334] "Generic (PLEG): container finished" podID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerID="3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4" exitCode=0 Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.516290 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fxdw" event={"ID":"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0","Type":"ContainerDied","Data":"3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4"} Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.516320 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9fxdw" event={"ID":"a1c49a74-f205-4feb-b0ed-4dd32c4e99e0","Type":"ContainerDied","Data":"bf2812e705cc64e7dd393e6ff7e16e4053f437f85fe0cfd07abe0ac45b2928cc"} Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.516340 4960 scope.go:117] "RemoveContainer" containerID="3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.516518 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9fxdw" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.547220 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fxdw"] Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.549068 4960 scope.go:117] "RemoveContainer" containerID="2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.556782 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9fxdw"] Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.589057 4960 scope.go:117] "RemoveContainer" containerID="4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.635717 4960 scope.go:117] "RemoveContainer" containerID="3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4" Oct 02 09:32:22 crc kubenswrapper[4960]: E1002 09:32:22.636603 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4\": container with ID starting with 3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4 not found: ID does not exist" containerID="3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.636644 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4"} err="failed to get container status \"3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4\": rpc error: code = NotFound desc = could not find container \"3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4\": container with ID starting with 3a3425a42f4b6d62bb324084a36f35c8acf7cd48b3f29f3a2b3177e0c3943ec4 not found: ID does not exist" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.636670 4960 scope.go:117] "RemoveContainer" containerID="2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7" Oct 02 09:32:22 crc kubenswrapper[4960]: E1002 09:32:22.637258 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7\": container with ID starting with 2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7 not found: ID does not exist" containerID="2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.637319 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7"} err="failed to get container status \"2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7\": rpc error: code = NotFound desc = could not find container \"2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7\": container with ID starting with 2038598cab97e7646e900edbb30d668f9a1144374b344b58281c71a8feeaf7c7 not found: ID does not exist" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.637364 4960 scope.go:117] "RemoveContainer" containerID="4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf" Oct 02 09:32:22 crc kubenswrapper[4960]: E1002 09:32:22.637773 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf\": container with ID starting with 4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf not found: ID does not exist" containerID="4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf" Oct 02 09:32:22 crc kubenswrapper[4960]: I1002 09:32:22.637810 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf"} err="failed to get container status \"4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf\": rpc error: code = NotFound desc = could not find container \"4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf\": container with ID starting with 4ead4a2d930a36f0cdcbaa8942b0d8976960327bc4852e6d9f3a34fd0b73d7cf not found: ID does not exist" Oct 02 09:32:24 crc kubenswrapper[4960]: I1002 09:32:24.342527 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" path="/var/lib/kubelet/pods/a1c49a74-f205-4feb-b0ed-4dd32c4e99e0/volumes" Oct 02 09:33:29 crc kubenswrapper[4960]: I1002 09:33:29.150473 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:33:29 crc kubenswrapper[4960]: I1002 09:33:29.151425 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:33:44 crc kubenswrapper[4960]: I1002 09:33:44.347999 4960 generic.go:334] "Generic (PLEG): container finished" podID="2b480863-210b-4fb7-8980-e939d5fc9603" containerID="3977b4bb2a3ade5145c826bb4a2a80db8dbc3508c6ae0461ad8b3624127c6098" exitCode=0 Oct 02 09:33:44 crc kubenswrapper[4960]: I1002 09:33:44.348840 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"2b480863-210b-4fb7-8980-e939d5fc9603","Type":"ContainerDied","Data":"3977b4bb2a3ade5145c826bb4a2a80db8dbc3508c6ae0461ad8b3624127c6098"} Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.783004 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.843725 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql2kx\" (UniqueName: \"kubernetes.io/projected/2b480863-210b-4fb7-8980-e939d5fc9603-kube-api-access-ql2kx\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844200 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-openstack-config-secret\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844242 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-temporary\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844304 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-workdir\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844401 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-clouds-config\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844429 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ceph\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844484 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-config\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844538 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844576 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-private-key\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844654 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-kubeconfig\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844694 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-public-key\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.844736 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ca-certs\") pod \"2b480863-210b-4fb7-8980-e939d5fc9603\" (UID: \"2b480863-210b-4fb7-8980-e939d5fc9603\") " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.846197 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.853674 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.853912 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ceph" (OuterVolumeSpecName: "ceph") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.867360 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b480863-210b-4fb7-8980-e939d5fc9603-kube-api-access-ql2kx" (OuterVolumeSpecName: "kube-api-access-ql2kx") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "kube-api-access-ql2kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.892033 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.896236 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.900837 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 02 09:33:45 crc kubenswrapper[4960]: E1002 09:33:45.901393 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b480863-210b-4fb7-8980-e939d5fc9603" containerName="tobiko-tests-tobiko" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.901415 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b480863-210b-4fb7-8980-e939d5fc9603" containerName="tobiko-tests-tobiko" Oct 02 09:33:45 crc kubenswrapper[4960]: E1002 09:33:45.901454 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="extract-utilities" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.901463 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="extract-utilities" Oct 02 09:33:45 crc kubenswrapper[4960]: E1002 09:33:45.901474 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="extract-content" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.901481 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="extract-content" Oct 02 09:33:45 crc kubenswrapper[4960]: E1002 09:33:45.901510 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="registry-server" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.901515 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="registry-server" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.901705 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b480863-210b-4fb7-8980-e939d5fc9603" containerName="tobiko-tests-tobiko" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.901734 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c49a74-f205-4feb-b0ed-4dd32c4e99e0" containerName="registry-server" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.902549 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.905673 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.911605 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.918914 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.924904 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.940888 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947339 4960 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947385 4960 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-kubeconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947396 4960 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947406 4960 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947415 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql2kx\" (UniqueName: \"kubernetes.io/projected/2b480863-210b-4fb7-8980-e939d5fc9603-kube-api-access-ql2kx\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947424 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947434 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947443 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2b480863-210b-4fb7-8980-e939d5fc9603-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947452 4960 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-tobiko-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.947498 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.952237 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:33:45 crc kubenswrapper[4960]: I1002 09:33:45.972288 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.049305 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.049608 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.049903 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050144 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050193 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050315 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050447 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050512 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050676 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050753 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwldd\" (UniqueName: \"kubernetes.io/projected/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kube-api-access-gwldd\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.050826 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.051037 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.051129 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.051890 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.109097 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154058 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154127 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154174 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154246 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154265 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154300 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154336 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154370 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154411 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwldd\" (UniqueName: \"kubernetes.io/projected/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kube-api-access-gwldd\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154433 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.154484 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.155958 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.156047 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.156329 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.157410 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.157927 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.158342 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.162055 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.162090 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.162368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.164853 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.176101 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwldd\" (UniqueName: \"kubernetes.io/projected/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kube-api-access-gwldd\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.351155 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.381418 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"2b480863-210b-4fb7-8980-e939d5fc9603","Type":"ContainerDied","Data":"e4c725a82405e230b8ccc0fe2ad4afbded74dfdaba5366b54ea32b672ac5fe06"} Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.381486 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4c725a82405e230b8ccc0fe2ad4afbded74dfdaba5366b54ea32b672ac5fe06" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.381591 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 02 09:33:46 crc kubenswrapper[4960]: I1002 09:33:46.906615 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 02 09:33:46 crc kubenswrapper[4960]: W1002 09:33:46.945848 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79e4f0e3_919f_44af_b6ae_9bd32f4ff5c8.slice/crio-d2d37c3b830768c5290a01d619990bfbccfc762d8707ce79cb2b2aaae4a49196 WatchSource:0}: Error finding container d2d37c3b830768c5290a01d619990bfbccfc762d8707ce79cb2b2aaae4a49196: Status 404 returned error can't find the container with id d2d37c3b830768c5290a01d619990bfbccfc762d8707ce79cb2b2aaae4a49196 Oct 02 09:33:47 crc kubenswrapper[4960]: I1002 09:33:47.401810 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8","Type":"ContainerStarted","Data":"d2d37c3b830768c5290a01d619990bfbccfc762d8707ce79cb2b2aaae4a49196"} Oct 02 09:33:47 crc kubenswrapper[4960]: I1002 09:33:47.505596 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "2b480863-210b-4fb7-8980-e939d5fc9603" (UID: "2b480863-210b-4fb7-8980-e939d5fc9603"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:33:47 crc kubenswrapper[4960]: I1002 09:33:47.596553 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2b480863-210b-4fb7-8980-e939d5fc9603-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:33:48 crc kubenswrapper[4960]: I1002 09:33:48.414151 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8","Type":"ContainerStarted","Data":"d5845c6da89a86b7f06afb705e8e242d76605223765baca1008314395c7f7225"} Oct 02 09:33:48 crc kubenswrapper[4960]: I1002 09:33:48.438429 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s01-sanity" podStartSLOduration=3.438366631 podStartE2EDuration="3.438366631s" podCreationTimestamp="2025-10-02 09:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:33:48.434934749 +0000 UTC m=+8249.466881056" watchObservedRunningTime="2025-10-02 09:33:48.438366631 +0000 UTC m=+8249.470312918" Oct 02 09:33:59 crc kubenswrapper[4960]: I1002 09:33:59.151020 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:33:59 crc kubenswrapper[4960]: I1002 09:33:59.152069 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.038255 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pbkwk"] Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.042059 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.059711 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pbkwk"] Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.187552 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-utilities\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.188097 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-catalog-content\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.188170 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvlfq\" (UniqueName: \"kubernetes.io/projected/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-kube-api-access-nvlfq\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.290463 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-utilities\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.290653 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-catalog-content\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.290695 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvlfq\" (UniqueName: \"kubernetes.io/projected/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-kube-api-access-nvlfq\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.291151 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-utilities\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.291243 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-catalog-content\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.311924 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvlfq\" (UniqueName: \"kubernetes.io/projected/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-kube-api-access-nvlfq\") pod \"certified-operators-pbkwk\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.364824 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:17 crc kubenswrapper[4960]: I1002 09:34:17.737619 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pbkwk"] Oct 02 09:34:18 crc kubenswrapper[4960]: I1002 09:34:18.714751 4960 generic.go:334] "Generic (PLEG): container finished" podID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerID="c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c" exitCode=0 Oct 02 09:34:18 crc kubenswrapper[4960]: I1002 09:34:18.714897 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbkwk" event={"ID":"9a28dfd7-94ab-4f2d-8912-2a91f46d408c","Type":"ContainerDied","Data":"c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c"} Oct 02 09:34:18 crc kubenswrapper[4960]: I1002 09:34:18.715380 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbkwk" event={"ID":"9a28dfd7-94ab-4f2d-8912-2a91f46d408c","Type":"ContainerStarted","Data":"9d36c3c4d86bdacb5678234646df3055d35b0adbbbca7e2d5fc3b28c1a303e7c"} Oct 02 09:34:20 crc kubenswrapper[4960]: I1002 09:34:20.754771 4960 generic.go:334] "Generic (PLEG): container finished" podID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerID="bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1" exitCode=0 Oct 02 09:34:20 crc kubenswrapper[4960]: I1002 09:34:20.756277 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbkwk" event={"ID":"9a28dfd7-94ab-4f2d-8912-2a91f46d408c","Type":"ContainerDied","Data":"bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1"} Oct 02 09:34:21 crc kubenswrapper[4960]: I1002 09:34:21.770600 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbkwk" event={"ID":"9a28dfd7-94ab-4f2d-8912-2a91f46d408c","Type":"ContainerStarted","Data":"4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67"} Oct 02 09:34:21 crc kubenswrapper[4960]: I1002 09:34:21.814374 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pbkwk" podStartSLOduration=2.249059761 podStartE2EDuration="4.814344158s" podCreationTimestamp="2025-10-02 09:34:17 +0000 UTC" firstStartedPulling="2025-10-02 09:34:18.717165668 +0000 UTC m=+8279.749111965" lastFinishedPulling="2025-10-02 09:34:21.282450075 +0000 UTC m=+8282.314396362" observedRunningTime="2025-10-02 09:34:21.804488096 +0000 UTC m=+8282.836434383" watchObservedRunningTime="2025-10-02 09:34:21.814344158 +0000 UTC m=+8282.846290455" Oct 02 09:34:27 crc kubenswrapper[4960]: I1002 09:34:27.365312 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:27 crc kubenswrapper[4960]: I1002 09:34:27.366096 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:27 crc kubenswrapper[4960]: I1002 09:34:27.424224 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:27 crc kubenswrapper[4960]: I1002 09:34:27.887178 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:27 crc kubenswrapper[4960]: I1002 09:34:27.941627 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pbkwk"] Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.150387 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.150811 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.150902 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.152215 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e3b98926708e1815f9439704caba866fca62e36285e3709b97655ae40054844"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.152317 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://3e3b98926708e1815f9439704caba866fca62e36285e3709b97655ae40054844" gracePeriod=600 Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.857411 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="3e3b98926708e1815f9439704caba866fca62e36285e3709b97655ae40054844" exitCode=0 Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.857544 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"3e3b98926708e1815f9439704caba866fca62e36285e3709b97655ae40054844"} Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.858090 4960 scope.go:117] "RemoveContainer" containerID="b08183f9a729e9113389216e0198bbca67e04bec1781ff185aeb30ee36d5dbc7" Oct 02 09:34:29 crc kubenswrapper[4960]: I1002 09:34:29.858191 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pbkwk" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="registry-server" containerID="cri-o://4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67" gracePeriod=2 Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.400373 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.510586 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-catalog-content\") pod \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.510963 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvlfq\" (UniqueName: \"kubernetes.io/projected/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-kube-api-access-nvlfq\") pod \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.511083 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-utilities\") pod \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\" (UID: \"9a28dfd7-94ab-4f2d-8912-2a91f46d408c\") " Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.512082 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-utilities" (OuterVolumeSpecName: "utilities") pod "9a28dfd7-94ab-4f2d-8912-2a91f46d408c" (UID: "9a28dfd7-94ab-4f2d-8912-2a91f46d408c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.520306 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-kube-api-access-nvlfq" (OuterVolumeSpecName: "kube-api-access-nvlfq") pod "9a28dfd7-94ab-4f2d-8912-2a91f46d408c" (UID: "9a28dfd7-94ab-4f2d-8912-2a91f46d408c"). InnerVolumeSpecName "kube-api-access-nvlfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.566259 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a28dfd7-94ab-4f2d-8912-2a91f46d408c" (UID: "9a28dfd7-94ab-4f2d-8912-2a91f46d408c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.614614 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.614653 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvlfq\" (UniqueName: \"kubernetes.io/projected/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-kube-api-access-nvlfq\") on node \"crc\" DevicePath \"\"" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.614663 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a28dfd7-94ab-4f2d-8912-2a91f46d408c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.869683 4960 generic.go:334] "Generic (PLEG): container finished" podID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerID="4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67" exitCode=0 Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.869783 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pbkwk" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.869779 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbkwk" event={"ID":"9a28dfd7-94ab-4f2d-8912-2a91f46d408c","Type":"ContainerDied","Data":"4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67"} Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.870101 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pbkwk" event={"ID":"9a28dfd7-94ab-4f2d-8912-2a91f46d408c","Type":"ContainerDied","Data":"9d36c3c4d86bdacb5678234646df3055d35b0adbbbca7e2d5fc3b28c1a303e7c"} Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.870127 4960 scope.go:117] "RemoveContainer" containerID="4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.872831 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442"} Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.901851 4960 scope.go:117] "RemoveContainer" containerID="bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.923354 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pbkwk"] Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.931170 4960 scope.go:117] "RemoveContainer" containerID="c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.932721 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pbkwk"] Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.992608 4960 scope.go:117] "RemoveContainer" containerID="4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67" Oct 02 09:34:30 crc kubenswrapper[4960]: E1002 09:34:30.994237 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67\": container with ID starting with 4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67 not found: ID does not exist" containerID="4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.994308 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67"} err="failed to get container status \"4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67\": rpc error: code = NotFound desc = could not find container \"4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67\": container with ID starting with 4246829b37d38fe324ec8053a5e365a4ab85e285036b0b9ac9744a7fb016ba67 not found: ID does not exist" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.994352 4960 scope.go:117] "RemoveContainer" containerID="bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1" Oct 02 09:34:30 crc kubenswrapper[4960]: E1002 09:34:30.995564 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1\": container with ID starting with bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1 not found: ID does not exist" containerID="bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.995616 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1"} err="failed to get container status \"bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1\": rpc error: code = NotFound desc = could not find container \"bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1\": container with ID starting with bde04dd06bc0ee5623bdd0b4af5faff26eb6bd6fd3f49c583bbe73bd7e008af1 not found: ID does not exist" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.995647 4960 scope.go:117] "RemoveContainer" containerID="c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c" Oct 02 09:34:30 crc kubenswrapper[4960]: E1002 09:34:30.996052 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c\": container with ID starting with c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c not found: ID does not exist" containerID="c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c" Oct 02 09:34:30 crc kubenswrapper[4960]: I1002 09:34:30.996081 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c"} err="failed to get container status \"c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c\": rpc error: code = NotFound desc = could not find container \"c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c\": container with ID starting with c04aa0af35a0e17a3e2e3a9f554ce7f3abaf83174a9572a102b5946c2193341c not found: ID does not exist" Oct 02 09:34:32 crc kubenswrapper[4960]: I1002 09:34:32.344625 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" path="/var/lib/kubelet/pods/9a28dfd7-94ab-4f2d-8912-2a91f46d408c/volumes" Oct 02 09:36:04 crc kubenswrapper[4960]: I1002 09:36:04.811252 4960 generic.go:334] "Generic (PLEG): container finished" podID="79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" containerID="d5845c6da89a86b7f06afb705e8e242d76605223765baca1008314395c7f7225" exitCode=0 Oct 02 09:36:04 crc kubenswrapper[4960]: I1002 09:36:04.811319 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8","Type":"ContainerDied","Data":"d5845c6da89a86b7f06afb705e8e242d76605223765baca1008314395c7f7225"} Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.313724 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.434813 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-clouds-config\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.434895 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ca-certs\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435038 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-public-key\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435077 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-config\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435122 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-workdir\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435198 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kubeconfig\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435262 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-temporary\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435307 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435359 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-openstack-config-secret\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435409 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwldd\" (UniqueName: \"kubernetes.io/projected/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kube-api-access-gwldd\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435471 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-private-key\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.435510 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ceph\") pod \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\" (UID: \"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8\") " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.438319 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.444514 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ceph" (OuterVolumeSpecName: "ceph") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.448096 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kube-api-access-gwldd" (OuterVolumeSpecName: "kube-api-access-gwldd") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "kube-api-access-gwldd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.460342 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.465763 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.472775 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.473190 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.478508 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.478653 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.505524 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.510428 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.538903 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.538941 4960 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.538956 4960 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.538966 4960 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.538998 4960 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kubeconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.539015 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.539048 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.539058 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.539072 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwldd\" (UniqueName: \"kubernetes.io/projected/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-kube-api-access-gwldd\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.539081 4960 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.539092 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.564521 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.641443 4960 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.836141 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8","Type":"ContainerDied","Data":"d2d37c3b830768c5290a01d619990bfbccfc762d8707ce79cb2b2aaae4a49196"} Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.836189 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2d37c3b830768c5290a01d619990bfbccfc762d8707ce79cb2b2aaae4a49196" Oct 02 09:36:06 crc kubenswrapper[4960]: I1002 09:36:06.836261 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 02 09:36:07 crc kubenswrapper[4960]: I1002 09:36:07.890108 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" (UID: "79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:36:07 crc kubenswrapper[4960]: I1002 09:36:07.979670 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.822186 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 02 09:36:08 crc kubenswrapper[4960]: E1002 09:36:08.822844 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="registry-server" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.822869 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="registry-server" Oct 02 09:36:08 crc kubenswrapper[4960]: E1002 09:36:08.822888 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="extract-content" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.822895 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="extract-content" Oct 02 09:36:08 crc kubenswrapper[4960]: E1002 09:36:08.822909 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" containerName="tobiko-tests-tobiko" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.822918 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" containerName="tobiko-tests-tobiko" Oct 02 09:36:08 crc kubenswrapper[4960]: E1002 09:36:08.822932 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="extract-utilities" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.822937 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="extract-utilities" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.823216 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8" containerName="tobiko-tests-tobiko" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.823259 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a28dfd7-94ab-4f2d-8912-2a91f46d408c" containerName="registry-server" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.824224 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:08 crc kubenswrapper[4960]: I1002 09:36:08.835810 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.002086 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/6213b3c4-9214-43d0-9ddf-782b3ec673af-kube-api-access-qcl86\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"6213b3c4-9214-43d0-9ddf-782b3ec673af\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.002311 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"6213b3c4-9214-43d0-9ddf-782b3ec673af\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.104598 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"6213b3c4-9214-43d0-9ddf-782b3ec673af\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.104736 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/6213b3c4-9214-43d0-9ddf-782b3ec673af-kube-api-access-qcl86\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"6213b3c4-9214-43d0-9ddf-782b3ec673af\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.105262 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"6213b3c4-9214-43d0-9ddf-782b3ec673af\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.129259 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/6213b3c4-9214-43d0-9ddf-782b3ec673af-kube-api-access-qcl86\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"6213b3c4-9214-43d0-9ddf-782b3ec673af\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.134664 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"6213b3c4-9214-43d0-9ddf-782b3ec673af\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.157547 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.639831 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.646886 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:36:09 crc kubenswrapper[4960]: I1002 09:36:09.879489 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"6213b3c4-9214-43d0-9ddf-782b3ec673af","Type":"ContainerStarted","Data":"dc7fe221a2eaa0865213afdc8fcef59c6cfcb35d5479f90e53437aaf8fabc146"} Oct 02 09:36:10 crc kubenswrapper[4960]: I1002 09:36:10.898706 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"6213b3c4-9214-43d0-9ddf-782b3ec673af","Type":"ContainerStarted","Data":"de741f77e516cfd6e8aa91bb7d9c973641d442219e6a6645a0d52e7d014a202f"} Oct 02 09:36:10 crc kubenswrapper[4960]: I1002 09:36:10.922892 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" podStartSLOduration=2.3696397129999998 podStartE2EDuration="2.922865893s" podCreationTimestamp="2025-10-02 09:36:08 +0000 UTC" firstStartedPulling="2025-10-02 09:36:09.646599028 +0000 UTC m=+8390.678545315" lastFinishedPulling="2025-10-02 09:36:10.199825218 +0000 UTC m=+8391.231771495" observedRunningTime="2025-10-02 09:36:10.91938732 +0000 UTC m=+8391.951333607" watchObservedRunningTime="2025-10-02 09:36:10.922865893 +0000 UTC m=+8391.954812180" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.513266 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.518600 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.520518 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.520909 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.530836 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.607962 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.608338 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.608481 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.608618 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.608735 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.608840 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.608989 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.609143 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ceph\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.609361 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.609556 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl6vx\" (UniqueName: \"kubernetes.io/projected/baea4578-7ae2-42d3-a07a-ef15c47c886b-kube-api-access-bl6vx\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710285 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710358 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710393 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710417 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710440 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710463 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ceph\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710596 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl6vx\" (UniqueName: \"kubernetes.io/projected/baea4578-7ae2-42d3-a07a-ef15c47c886b-kube-api-access-bl6vx\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710650 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710683 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.710754 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.711051 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.712159 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.712523 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.716694 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ceph\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.717018 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.717611 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.718474 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.718660 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.729769 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl6vx\" (UniqueName: \"kubernetes.io/projected/baea4578-7ae2-42d3-a07a-ef15c47c886b-kube-api-access-bl6vx\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.744325 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ansibletest-ansibletest\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " pod="openstack/ansibletest-ansibletest" Oct 02 09:36:30 crc kubenswrapper[4960]: I1002 09:36:30.848966 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 02 09:36:31 crc kubenswrapper[4960]: I1002 09:36:31.321477 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 02 09:36:32 crc kubenswrapper[4960]: I1002 09:36:32.104299 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"baea4578-7ae2-42d3-a07a-ef15c47c886b","Type":"ContainerStarted","Data":"119ee838a926aa1b38a2927e9532ba3794691939fd8985d53d3918ca1d42b998"} Oct 02 09:36:58 crc kubenswrapper[4960]: E1002 09:36:58.779150 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified" Oct 02 09:36:58 crc kubenswrapper[4960]: E1002 09:36:58.780098 4960 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 02 09:36:58 crc kubenswrapper[4960]: container &Container{Name:ansibletest-ansibletest,Image:quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_ANSIBLE_EXTRA_VARS,Value:-e manual_run=false,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_FILE_EXTRA_VARS,Value:--- Oct 02 09:36:58 crc kubenswrapper[4960]: foo: bar Oct 02 09:36:58 crc kubenswrapper[4960]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_BRANCH,Value:,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_REPO,Value:https://github.com/ansible/test-playbooks,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_INVENTORY,Value:localhost ansible_connection=local ansible_python_interpreter=python3 Oct 02 09:36:58 crc kubenswrapper[4960]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_PLAYBOOK,Value:./debug.yml,ValueFrom:nil,},EnvVar{Name:POD_DEBUG,Value:false,ValueFrom:nil,},EnvVar{Name:POD_INSTALL_COLLECTIONS,Value:,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/ansible,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/AnsibleTests/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/ansible/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/var/lib/ansible/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:workload-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/test_keypair.key,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:compute-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/.ssh/compute_id,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bl6vx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*227,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*227,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ansibletest-ansibletest_openstack(baea4578-7ae2-42d3-a07a-ef15c47c886b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 02 09:36:58 crc kubenswrapper[4960]: > logger="UnhandledError" Oct 02 09:36:58 crc kubenswrapper[4960]: E1002 09:36:58.781524 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ansibletest-ansibletest" podUID="baea4578-7ae2-42d3-a07a-ef15c47c886b" Oct 02 09:36:59 crc kubenswrapper[4960]: I1002 09:36:59.150569 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:36:59 crc kubenswrapper[4960]: I1002 09:36:59.150649 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:36:59 crc kubenswrapper[4960]: E1002 09:36:59.422329 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified\\\"\"" pod="openstack/ansibletest-ansibletest" podUID="baea4578-7ae2-42d3-a07a-ef15c47c886b" Oct 02 09:37:16 crc kubenswrapper[4960]: I1002 09:37:16.595334 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"baea4578-7ae2-42d3-a07a-ef15c47c886b","Type":"ContainerStarted","Data":"e494e033eaf375447f90a85f64e9b11df5c2077ee37b508d7fe4ad2fa565cf3d"} Oct 02 09:37:16 crc kubenswrapper[4960]: I1002 09:37:16.627642 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ansibletest-ansibletest" podStartSLOduration=4.040348913 podStartE2EDuration="47.627608945s" podCreationTimestamp="2025-10-02 09:36:29 +0000 UTC" firstStartedPulling="2025-10-02 09:36:31.330350407 +0000 UTC m=+8412.362296694" lastFinishedPulling="2025-10-02 09:37:14.917610439 +0000 UTC m=+8455.949556726" observedRunningTime="2025-10-02 09:37:16.620602899 +0000 UTC m=+8457.652549196" watchObservedRunningTime="2025-10-02 09:37:16.627608945 +0000 UTC m=+8457.659555232" Oct 02 09:37:20 crc kubenswrapper[4960]: I1002 09:37:20.638928 4960 generic.go:334] "Generic (PLEG): container finished" podID="baea4578-7ae2-42d3-a07a-ef15c47c886b" containerID="e494e033eaf375447f90a85f64e9b11df5c2077ee37b508d7fe4ad2fa565cf3d" exitCode=0 Oct 02 09:37:20 crc kubenswrapper[4960]: I1002 09:37:20.639210 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"baea4578-7ae2-42d3-a07a-ef15c47c886b","Type":"ContainerDied","Data":"e494e033eaf375447f90a85f64e9b11df5c2077ee37b508d7fe4ad2fa565cf3d"} Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.008640 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.181405 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-workdir\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.182070 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config-secret\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.182201 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl6vx\" (UniqueName: \"kubernetes.io/projected/baea4578-7ae2-42d3-a07a-ef15c47c886b-kube-api-access-bl6vx\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.182315 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.183147 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ceph\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.183251 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-workload-ssh-secret\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.183647 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-compute-ssh-secret\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.183695 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ca-certs\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.183841 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.183948 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-temporary\") pod \"baea4578-7ae2-42d3-a07a-ef15c47c886b\" (UID: \"baea4578-7ae2-42d3-a07a-ef15c47c886b\") " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.185604 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.191402 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.191560 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ceph" (OuterVolumeSpecName: "ceph") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.194281 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baea4578-7ae2-42d3-a07a-ef15c47c886b-kube-api-access-bl6vx" (OuterVolumeSpecName: "kube-api-access-bl6vx") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "kube-api-access-bl6vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.197355 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.224323 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-compute-ssh-secret" (OuterVolumeSpecName: "compute-ssh-secret") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "compute-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.228170 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.239260 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-workload-ssh-secret" (OuterVolumeSpecName: "workload-ssh-secret") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "workload-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.251915 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.255737 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "baea4578-7ae2-42d3-a07a-ef15c47c886b" (UID: "baea4578-7ae2-42d3-a07a-ef15c47c886b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287266 4960 reconciler_common.go:293] "Volume detached for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-compute-ssh-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287322 4960 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287363 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287377 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287394 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/baea4578-7ae2-42d3-a07a-ef15c47c886b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287405 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287416 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl6vx\" (UniqueName: \"kubernetes.io/projected/baea4578-7ae2-42d3-a07a-ef15c47c886b-kube-api-access-bl6vx\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287424 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/baea4578-7ae2-42d3-a07a-ef15c47c886b-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287433 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.287442 4960 reconciler_common.go:293] "Volume detached for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/baea4578-7ae2-42d3-a07a-ef15c47c886b-workload-ssh-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.313751 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.389915 4960 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.662641 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"baea4578-7ae2-42d3-a07a-ef15c47c886b","Type":"ContainerDied","Data":"119ee838a926aa1b38a2927e9532ba3794691939fd8985d53d3918ca1d42b998"} Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.662701 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="119ee838a926aa1b38a2927e9532ba3794691939fd8985d53d3918ca1d42b998" Oct 02 09:37:22 crc kubenswrapper[4960]: I1002 09:37:22.663041 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 02 09:37:26 crc kubenswrapper[4960]: I1002 09:37:26.735523 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 02 09:37:26 crc kubenswrapper[4960]: E1002 09:37:26.736509 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baea4578-7ae2-42d3-a07a-ef15c47c886b" containerName="ansibletest-ansibletest" Oct 02 09:37:26 crc kubenswrapper[4960]: I1002 09:37:26.736528 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="baea4578-7ae2-42d3-a07a-ef15c47c886b" containerName="ansibletest-ansibletest" Oct 02 09:37:26 crc kubenswrapper[4960]: I1002 09:37:26.736792 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="baea4578-7ae2-42d3-a07a-ef15c47c886b" containerName="ansibletest-ansibletest" Oct 02 09:37:26 crc kubenswrapper[4960]: I1002 09:37:26.737766 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:26 crc kubenswrapper[4960]: I1002 09:37:26.749483 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 02 09:37:26 crc kubenswrapper[4960]: I1002 09:37:26.905205 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pnt4\" (UniqueName: \"kubernetes.io/projected/a15540db-b8e9-444e-9d55-7bcb2b1c4531-kube-api-access-5pnt4\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"a15540db-b8e9-444e-9d55-7bcb2b1c4531\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:26 crc kubenswrapper[4960]: I1002 09:37:26.905927 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"a15540db-b8e9-444e-9d55-7bcb2b1c4531\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.012210 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pnt4\" (UniqueName: \"kubernetes.io/projected/a15540db-b8e9-444e-9d55-7bcb2b1c4531-kube-api-access-5pnt4\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"a15540db-b8e9-444e-9d55-7bcb2b1c4531\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.012371 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"a15540db-b8e9-444e-9d55-7bcb2b1c4531\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.013231 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"a15540db-b8e9-444e-9d55-7bcb2b1c4531\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.038845 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pnt4\" (UniqueName: \"kubernetes.io/projected/a15540db-b8e9-444e-9d55-7bcb2b1c4531-kube-api-access-5pnt4\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"a15540db-b8e9-444e-9d55-7bcb2b1c4531\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.041858 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"a15540db-b8e9-444e-9d55-7bcb2b1c4531\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.065510 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.542119 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 02 09:37:27 crc kubenswrapper[4960]: I1002 09:37:27.721356 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"a15540db-b8e9-444e-9d55-7bcb2b1c4531","Type":"ContainerStarted","Data":"e2f2dfa1fa45cd76a795d385d7b0a24578cdb31e4496e0a82acf974fd47ea12a"} Oct 02 09:37:28 crc kubenswrapper[4960]: I1002 09:37:28.731808 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"a15540db-b8e9-444e-9d55-7bcb2b1c4531","Type":"ContainerStarted","Data":"e791ffcbed74bbb5659e58d7836d539fdea49c6d9b6583643c248adbc0b4fff8"} Oct 02 09:37:28 crc kubenswrapper[4960]: I1002 09:37:28.756378 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" podStartSLOduration=2.087975779 podStartE2EDuration="2.756350621s" podCreationTimestamp="2025-10-02 09:37:26 +0000 UTC" firstStartedPulling="2025-10-02 09:37:27.543587955 +0000 UTC m=+8468.575534242" lastFinishedPulling="2025-10-02 09:37:28.211962797 +0000 UTC m=+8469.243909084" observedRunningTime="2025-10-02 09:37:28.747582918 +0000 UTC m=+8469.779529215" watchObservedRunningTime="2025-10-02 09:37:28.756350621 +0000 UTC m=+8469.788296908" Oct 02 09:37:29 crc kubenswrapper[4960]: I1002 09:37:29.149989 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:37:29 crc kubenswrapper[4960]: I1002 09:37:29.150080 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.382634 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.386085 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.390716 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.390916 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizontest-tests-horizontesthorizontest-config" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.398867 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478204 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478254 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478282 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqjfl\" (UniqueName: \"kubernetes.io/projected/179d0cf9-edfa-4b98-8708-1503790d862e-kube-api-access-pqjfl\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478371 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478399 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478443 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478478 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.478499 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580052 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580152 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580176 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580301 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580319 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580346 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqjfl\" (UniqueName: \"kubernetes.io/projected/179d0cf9-edfa-4b98-8708-1503790d862e-kube-api-access-pqjfl\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580389 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580410 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.580682 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.581649 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.581747 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.581759 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.588527 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.589934 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.598372 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.607854 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqjfl\" (UniqueName: \"kubernetes.io/projected/179d0cf9-edfa-4b98-8708-1503790d862e-kube-api-access-pqjfl\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.617891 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:46 crc kubenswrapper[4960]: I1002 09:37:46.709917 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 02 09:37:47 crc kubenswrapper[4960]: I1002 09:37:47.180938 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 02 09:37:47 crc kubenswrapper[4960]: I1002 09:37:47.955139 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"179d0cf9-edfa-4b98-8708-1503790d862e","Type":"ContainerStarted","Data":"e08fb737db88f92d336d0f25fba23ab46dc4f105121454c4f8ed57184c8b411d"} Oct 02 09:37:59 crc kubenswrapper[4960]: I1002 09:37:59.149815 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:37:59 crc kubenswrapper[4960]: I1002 09:37:59.152143 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:37:59 crc kubenswrapper[4960]: I1002 09:37:59.152317 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:37:59 crc kubenswrapper[4960]: I1002 09:37:59.153561 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:37:59 crc kubenswrapper[4960]: I1002 09:37:59.153752 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" gracePeriod=600 Oct 02 09:38:00 crc kubenswrapper[4960]: E1002 09:38:00.075380 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:38:00 crc kubenswrapper[4960]: I1002 09:38:00.099917 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" exitCode=0 Oct 02 09:38:00 crc kubenswrapper[4960]: I1002 09:38:00.099989 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442"} Oct 02 09:38:00 crc kubenswrapper[4960]: I1002 09:38:00.100040 4960 scope.go:117] "RemoveContainer" containerID="3e3b98926708e1815f9439704caba866fca62e36285e3709b97655ae40054844" Oct 02 09:38:00 crc kubenswrapper[4960]: I1002 09:38:00.100683 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:38:00 crc kubenswrapper[4960]: E1002 09:38:00.101060 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:38:11 crc kubenswrapper[4960]: I1002 09:38:11.330238 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:38:11 crc kubenswrapper[4960]: E1002 09:38:11.331391 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:38:22 crc kubenswrapper[4960]: E1002 09:38:22.469779 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizontest:current-podified" Oct 02 09:38:22 crc kubenswrapper[4960]: E1002 09:38:22.470635 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizontest-tests-horizontest,Image:quay.io/podified-antelope-centos9/openstack-horizontest:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADMIN_PASSWORD,Value:12345678,ValueFrom:nil,},EnvVar{Name:ADMIN_USERNAME,Value:admin,ValueFrom:nil,},EnvVar{Name:AUTH_URL,Value:https://keystone-public-openstack.apps-crc.testing,ValueFrom:nil,},EnvVar{Name:DASHBOARD_URL,Value:https://horizon-openstack.apps-crc.testing/,ValueFrom:nil,},EnvVar{Name:EXTRA_FLAG,Value:not pagination and test_users.py,ValueFrom:nil,},EnvVar{Name:FLAVOR_NAME,Value:m1.tiny,ValueFrom:nil,},EnvVar{Name:HORIZONTEST_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:HORIZON_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:HORIZON_LOGS_DIR_NAME,Value:horizon,ValueFrom:nil,},EnvVar{Name:HORIZON_REPO_BRANCH,Value:master,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE,Value:/var/lib/horizontest/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE_NAME,Value:cirros-0.6.2-x86_64-disk,ValueFrom:nil,},EnvVar{Name:IMAGE_URL,Value:http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:PASSWORD,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME_XPATH,Value://*[@class=\"context-project\"]//ancestor::ul,ValueFrom:nil,},EnvVar{Name:REPO_URL,Value:https://review.opendev.org/openstack/horizon,ValueFrom:nil,},EnvVar{Name:USER_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/horizontest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/horizontest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/horizontest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pqjfl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42455,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42455,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizontest-tests-horizontest_openstack(179d0cf9-edfa-4b98-8708-1503790d862e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:38:22 crc kubenswrapper[4960]: E1002 09:38:22.472194 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizontest-tests-horizontest" podUID="179d0cf9-edfa-4b98-8708-1503790d862e" Oct 02 09:38:23 crc kubenswrapper[4960]: E1002 09:38:23.428240 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizontest:current-podified\\\"\"" pod="openstack/horizontest-tests-horizontest" podUID="179d0cf9-edfa-4b98-8708-1503790d862e" Oct 02 09:38:25 crc kubenswrapper[4960]: I1002 09:38:25.329876 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:38:25 crc kubenswrapper[4960]: E1002 09:38:25.330532 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:38:40 crc kubenswrapper[4960]: I1002 09:38:40.337254 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:38:40 crc kubenswrapper[4960]: E1002 09:38:40.338339 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:38:40 crc kubenswrapper[4960]: I1002 09:38:40.621573 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"179d0cf9-edfa-4b98-8708-1503790d862e","Type":"ContainerStarted","Data":"9556c93aaa50b30c6a14e1563eae7686cce5aba6ab103217958642584dc32658"} Oct 02 09:38:40 crc kubenswrapper[4960]: I1002 09:38:40.662933 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizontest-tests-horizontest" podStartSLOduration=3.362505071 podStartE2EDuration="55.662903421s" podCreationTimestamp="2025-10-02 09:37:45 +0000 UTC" firstStartedPulling="2025-10-02 09:37:47.195120281 +0000 UTC m=+8488.227066568" lastFinishedPulling="2025-10-02 09:38:39.495518631 +0000 UTC m=+8540.527464918" observedRunningTime="2025-10-02 09:38:40.652575476 +0000 UTC m=+8541.684521763" watchObservedRunningTime="2025-10-02 09:38:40.662903421 +0000 UTC m=+8541.694849708" Oct 02 09:38:52 crc kubenswrapper[4960]: I1002 09:38:52.330816 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:38:52 crc kubenswrapper[4960]: E1002 09:38:52.332195 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:39:03 crc kubenswrapper[4960]: I1002 09:39:03.330395 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:39:03 crc kubenswrapper[4960]: E1002 09:39:03.331511 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:39:18 crc kubenswrapper[4960]: I1002 09:39:18.331238 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:39:18 crc kubenswrapper[4960]: E1002 09:39:18.332472 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:39:29 crc kubenswrapper[4960]: I1002 09:39:29.330533 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:39:29 crc kubenswrapper[4960]: E1002 09:39:29.331436 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:39:41 crc kubenswrapper[4960]: I1002 09:39:41.331110 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:39:41 crc kubenswrapper[4960]: E1002 09:39:41.332074 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:39:55 crc kubenswrapper[4960]: I1002 09:39:55.330087 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:39:55 crc kubenswrapper[4960]: E1002 09:39:55.330831 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:40:08 crc kubenswrapper[4960]: I1002 09:40:08.330439 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:40:08 crc kubenswrapper[4960]: E1002 09:40:08.331468 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:40:23 crc kubenswrapper[4960]: I1002 09:40:23.330421 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:40:23 crc kubenswrapper[4960]: E1002 09:40:23.331546 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:40:37 crc kubenswrapper[4960]: I1002 09:40:37.331054 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:40:37 crc kubenswrapper[4960]: E1002 09:40:37.332232 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:40:41 crc kubenswrapper[4960]: I1002 09:40:41.973148 4960 generic.go:334] "Generic (PLEG): container finished" podID="179d0cf9-edfa-4b98-8708-1503790d862e" containerID="9556c93aaa50b30c6a14e1563eae7686cce5aba6ab103217958642584dc32658" exitCode=0 Oct 02 09:40:41 crc kubenswrapper[4960]: I1002 09:40:41.973269 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"179d0cf9-edfa-4b98-8708-1503790d862e","Type":"ContainerDied","Data":"9556c93aaa50b30c6a14e1563eae7686cce5aba6ab103217958642584dc32658"} Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.482124 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.609430 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ca-certs\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.609552 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-temporary\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.609663 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-workdir\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.609762 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-openstack-config-secret\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.609804 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-clouds-config\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.609886 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ceph\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.610074 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqjfl\" (UniqueName: \"kubernetes.io/projected/179d0cf9-edfa-4b98-8708-1503790d862e-kube-api-access-pqjfl\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.610108 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"179d0cf9-edfa-4b98-8708-1503790d862e\" (UID: \"179d0cf9-edfa-4b98-8708-1503790d862e\") " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.610591 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.611326 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.632786 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/179d0cf9-edfa-4b98-8708-1503790d862e-kube-api-access-pqjfl" (OuterVolumeSpecName: "kube-api-access-pqjfl") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "kube-api-access-pqjfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.632843 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ceph" (OuterVolumeSpecName: "ceph") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.640169 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.658399 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.684135 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.688248 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.713397 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.713508 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.713549 4960 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.713565 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqjfl\" (UniqueName: \"kubernetes.io/projected/179d0cf9-edfa-4b98-8708-1503790d862e-kube-api-access-pqjfl\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.713628 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.713642 4960 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/179d0cf9-edfa-4b98-8708-1503790d862e-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.740053 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.814983 4960 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.881438 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "179d0cf9-edfa-4b98-8708-1503790d862e" (UID: "179d0cf9-edfa-4b98-8708-1503790d862e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.917005 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/179d0cf9-edfa-4b98-8708-1503790d862e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.994824 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"179d0cf9-edfa-4b98-8708-1503790d862e","Type":"ContainerDied","Data":"e08fb737db88f92d336d0f25fba23ab46dc4f105121454c4f8ed57184c8b411d"} Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.994874 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e08fb737db88f92d336d0f25fba23ab46dc4f105121454c4f8ed57184c8b411d" Oct 02 09:40:43 crc kubenswrapper[4960]: I1002 09:40:43.994879 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.309206 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 02 09:40:51 crc kubenswrapper[4960]: E1002 09:40:51.311481 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179d0cf9-edfa-4b98-8708-1503790d862e" containerName="horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.311544 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="179d0cf9-edfa-4b98-8708-1503790d862e" containerName="horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.313020 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="179d0cf9-edfa-4b98-8708-1503790d862e" containerName="horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.314086 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.318527 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.391658 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh8dm\" (UniqueName: \"kubernetes.io/projected/d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f-kube-api-access-jh8dm\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.392248 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.494727 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh8dm\" (UniqueName: \"kubernetes.io/projected/d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f-kube-api-access-jh8dm\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.494922 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.495482 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.516372 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh8dm\" (UniqueName: \"kubernetes.io/projected/d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f-kube-api-access-jh8dm\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.524438 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: I1002 09:40:51.648444 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 02 09:40:51 crc kubenswrapper[4960]: E1002 09:40:51.648566 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:40:52 crc kubenswrapper[4960]: I1002 09:40:52.151327 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 02 09:40:52 crc kubenswrapper[4960]: E1002 09:40:52.160354 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:40:52 crc kubenswrapper[4960]: I1002 09:40:52.331442 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:40:52 crc kubenswrapper[4960]: E1002 09:40:52.334238 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:40:53 crc kubenswrapper[4960]: I1002 09:40:53.098210 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f","Type":"ContainerStarted","Data":"7ad0f1e771f5929dce9add50dbe559e20651aa595e7a67d5d29ce5100a3a871c"} Oct 02 09:40:53 crc kubenswrapper[4960]: E1002 09:40:53.183274 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:40:54 crc kubenswrapper[4960]: I1002 09:40:54.111622 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f","Type":"ContainerStarted","Data":"990f4fa6ee03132c0fbb093b85426c53dc750571b13231ac99fa3fd62ea01174"} Oct 02 09:40:54 crc kubenswrapper[4960]: E1002 09:40:54.113243 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:40:55 crc kubenswrapper[4960]: E1002 09:40:55.124313 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:41:05 crc kubenswrapper[4960]: I1002 09:41:05.331137 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:41:05 crc kubenswrapper[4960]: E1002 09:41:05.331942 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.676646 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" podStartSLOduration=21.655293484 podStartE2EDuration="22.67662161s" podCreationTimestamp="2025-10-02 09:40:51 +0000 UTC" firstStartedPulling="2025-10-02 09:40:52.161826182 +0000 UTC m=+8673.193772469" lastFinishedPulling="2025-10-02 09:40:53.183154308 +0000 UTC m=+8674.215100595" observedRunningTime="2025-10-02 09:40:54.127468536 +0000 UTC m=+8675.159414823" watchObservedRunningTime="2025-10-02 09:41:13.67662161 +0000 UTC m=+8694.708567897" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.681568 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2k8qc/must-gather-ltjpx"] Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.684242 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.686579 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2k8qc"/"openshift-service-ca.crt" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.686799 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2k8qc"/"default-dockercfg-bdqr9" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.695773 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2k8qc"/"kube-root-ca.crt" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.699221 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2k8qc/must-gather-ltjpx"] Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.787520 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75cbd459-1ded-44ea-94bf-74218922d643-must-gather-output\") pod \"must-gather-ltjpx\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.787959 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2n86\" (UniqueName: \"kubernetes.io/projected/75cbd459-1ded-44ea-94bf-74218922d643-kube-api-access-b2n86\") pod \"must-gather-ltjpx\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.890372 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2n86\" (UniqueName: \"kubernetes.io/projected/75cbd459-1ded-44ea-94bf-74218922d643-kube-api-access-b2n86\") pod \"must-gather-ltjpx\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.890640 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75cbd459-1ded-44ea-94bf-74218922d643-must-gather-output\") pod \"must-gather-ltjpx\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.891256 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75cbd459-1ded-44ea-94bf-74218922d643-must-gather-output\") pod \"must-gather-ltjpx\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:13 crc kubenswrapper[4960]: I1002 09:41:13.925961 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2n86\" (UniqueName: \"kubernetes.io/projected/75cbd459-1ded-44ea-94bf-74218922d643-kube-api-access-b2n86\") pod \"must-gather-ltjpx\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:14 crc kubenswrapper[4960]: I1002 09:41:14.007084 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:41:14 crc kubenswrapper[4960]: I1002 09:41:14.481927 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2k8qc/must-gather-ltjpx"] Oct 02 09:41:14 crc kubenswrapper[4960]: I1002 09:41:14.504573 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:41:15 crc kubenswrapper[4960]: I1002 09:41:15.332940 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" event={"ID":"75cbd459-1ded-44ea-94bf-74218922d643","Type":"ContainerStarted","Data":"8b9d761ca9304814d291a064de3772ed8c3045f61d49ceb238138eb1843e83ed"} Oct 02 09:41:16 crc kubenswrapper[4960]: I1002 09:41:16.331924 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:41:16 crc kubenswrapper[4960]: E1002 09:41:16.332655 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:41:22 crc kubenswrapper[4960]: I1002 09:41:22.414510 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" event={"ID":"75cbd459-1ded-44ea-94bf-74218922d643","Type":"ContainerStarted","Data":"2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de"} Oct 02 09:41:22 crc kubenswrapper[4960]: I1002 09:41:22.415794 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" event={"ID":"75cbd459-1ded-44ea-94bf-74218922d643","Type":"ContainerStarted","Data":"d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d"} Oct 02 09:41:22 crc kubenswrapper[4960]: I1002 09:41:22.439910 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" podStartSLOduration=2.55791348 podStartE2EDuration="9.439888482s" podCreationTimestamp="2025-10-02 09:41:13 +0000 UTC" firstStartedPulling="2025-10-02 09:41:14.504292816 +0000 UTC m=+8695.536239103" lastFinishedPulling="2025-10-02 09:41:21.386267818 +0000 UTC m=+8702.418214105" observedRunningTime="2025-10-02 09:41:22.436096961 +0000 UTC m=+8703.468043268" watchObservedRunningTime="2025-10-02 09:41:22.439888482 +0000 UTC m=+8703.471834769" Oct 02 09:41:29 crc kubenswrapper[4960]: I1002 09:41:29.330748 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:41:29 crc kubenswrapper[4960]: E1002 09:41:29.332047 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:41:30 crc kubenswrapper[4960]: E1002 09:41:30.605458 4960 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.177:57446->38.102.83.177:37215: write tcp 38.102.83.177:57446->38.102.83.177:37215: write: broken pipe Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.223142 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-4h66m"] Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.224921 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.325968 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-host\") pod \"crc-debug-4h66m\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.326283 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6t26\" (UniqueName: \"kubernetes.io/projected/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-kube-api-access-z6t26\") pod \"crc-debug-4h66m\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.428385 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-host\") pod \"crc-debug-4h66m\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.428533 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6t26\" (UniqueName: \"kubernetes.io/projected/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-kube-api-access-z6t26\") pod \"crc-debug-4h66m\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.428733 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-host\") pod \"crc-debug-4h66m\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.455585 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6t26\" (UniqueName: \"kubernetes.io/projected/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-kube-api-access-z6t26\") pod \"crc-debug-4h66m\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:31 crc kubenswrapper[4960]: I1002 09:41:31.552001 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:41:32 crc kubenswrapper[4960]: I1002 09:41:32.522845 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" event={"ID":"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c","Type":"ContainerStarted","Data":"b0c7f939bf608bcc33ab8e071a91356fdd971b181c067ad9ebf86f262e8d000e"} Oct 02 09:41:43 crc kubenswrapper[4960]: I1002 09:41:43.330750 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:41:43 crc kubenswrapper[4960]: E1002 09:41:43.331876 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:41:45 crc kubenswrapper[4960]: I1002 09:41:45.685740 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" event={"ID":"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c","Type":"ContainerStarted","Data":"1bb393d2c30f15aca66d8d4163d2af2f3f5c21f8c48ba5eba8e60b8a5a382d68"} Oct 02 09:41:45 crc kubenswrapper[4960]: I1002 09:41:45.703630 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" podStartSLOduration=1.671459813 podStartE2EDuration="14.703608139s" podCreationTimestamp="2025-10-02 09:41:31 +0000 UTC" firstStartedPulling="2025-10-02 09:41:31.622493783 +0000 UTC m=+8712.654440070" lastFinishedPulling="2025-10-02 09:41:44.654642109 +0000 UTC m=+8725.686588396" observedRunningTime="2025-10-02 09:41:45.700659761 +0000 UTC m=+8726.732606048" watchObservedRunningTime="2025-10-02 09:41:45.703608139 +0000 UTC m=+8726.735554426" Oct 02 09:41:54 crc kubenswrapper[4960]: I1002 09:41:54.333592 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:41:54 crc kubenswrapper[4960]: E1002 09:41:54.334709 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:42:05 crc kubenswrapper[4960]: I1002 09:42:05.330685 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:42:05 crc kubenswrapper[4960]: E1002 09:42:05.331521 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:42:20 crc kubenswrapper[4960]: I1002 09:42:20.331311 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:42:20 crc kubenswrapper[4960]: E1002 09:42:20.331428 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:42:20 crc kubenswrapper[4960]: E1002 09:42:20.333648 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:42:26 crc kubenswrapper[4960]: I1002 09:42:26.792620 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gkw7q"] Oct 02 09:42:26 crc kubenswrapper[4960]: I1002 09:42:26.795559 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:26 crc kubenswrapper[4960]: I1002 09:42:26.815379 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkw7q"] Oct 02 09:42:26 crc kubenswrapper[4960]: I1002 09:42:26.954838 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-catalog-content\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:26 crc kubenswrapper[4960]: I1002 09:42:26.955442 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-utilities\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:26 crc kubenswrapper[4960]: I1002 09:42:26.955739 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7442\" (UniqueName: \"kubernetes.io/projected/56df1e3a-f004-4228-ab06-3020e74d27b9-kube-api-access-r7442\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.057578 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7442\" (UniqueName: \"kubernetes.io/projected/56df1e3a-f004-4228-ab06-3020e74d27b9-kube-api-access-r7442\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.057818 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-catalog-content\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.057856 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-utilities\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.058486 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-utilities\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.059276 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-catalog-content\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.096159 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7442\" (UniqueName: \"kubernetes.io/projected/56df1e3a-f004-4228-ab06-3020e74d27b9-kube-api-access-r7442\") pod \"redhat-marketplace-gkw7q\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.119290 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:27 crc kubenswrapper[4960]: I1002 09:42:27.708687 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkw7q"] Oct 02 09:42:28 crc kubenswrapper[4960]: I1002 09:42:28.196772 4960 generic.go:334] "Generic (PLEG): container finished" podID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerID="a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff" exitCode=0 Oct 02 09:42:28 crc kubenswrapper[4960]: I1002 09:42:28.196853 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkw7q" event={"ID":"56df1e3a-f004-4228-ab06-3020e74d27b9","Type":"ContainerDied","Data":"a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff"} Oct 02 09:42:28 crc kubenswrapper[4960]: I1002 09:42:28.197528 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkw7q" event={"ID":"56df1e3a-f004-4228-ab06-3020e74d27b9","Type":"ContainerStarted","Data":"76cd8d7ec04edfaa1525ff756d0681cab1541e390c0e55124b6148c48790c034"} Oct 02 09:42:29 crc kubenswrapper[4960]: I1002 09:42:29.211484 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkw7q" event={"ID":"56df1e3a-f004-4228-ab06-3020e74d27b9","Type":"ContainerStarted","Data":"20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256"} Oct 02 09:42:31 crc kubenswrapper[4960]: I1002 09:42:31.330674 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:42:31 crc kubenswrapper[4960]: E1002 09:42:31.332089 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:42:34 crc kubenswrapper[4960]: I1002 09:42:34.272584 4960 generic.go:334] "Generic (PLEG): container finished" podID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerID="20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256" exitCode=0 Oct 02 09:42:34 crc kubenswrapper[4960]: I1002 09:42:34.272797 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkw7q" event={"ID":"56df1e3a-f004-4228-ab06-3020e74d27b9","Type":"ContainerDied","Data":"20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256"} Oct 02 09:42:36 crc kubenswrapper[4960]: I1002 09:42:36.293667 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkw7q" event={"ID":"56df1e3a-f004-4228-ab06-3020e74d27b9","Type":"ContainerStarted","Data":"a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf"} Oct 02 09:42:36 crc kubenswrapper[4960]: I1002 09:42:36.328589 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gkw7q" podStartSLOduration=2.724433069 podStartE2EDuration="10.328568132s" podCreationTimestamp="2025-10-02 09:42:26 +0000 UTC" firstStartedPulling="2025-10-02 09:42:28.19942384 +0000 UTC m=+8769.231370127" lastFinishedPulling="2025-10-02 09:42:35.803558903 +0000 UTC m=+8776.835505190" observedRunningTime="2025-10-02 09:42:36.313529133 +0000 UTC m=+8777.345475420" watchObservedRunningTime="2025-10-02 09:42:36.328568132 +0000 UTC m=+8777.360514419" Oct 02 09:42:37 crc kubenswrapper[4960]: I1002 09:42:37.120895 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:37 crc kubenswrapper[4960]: I1002 09:42:37.121637 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:37 crc kubenswrapper[4960]: I1002 09:42:37.185867 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:41 crc kubenswrapper[4960]: I1002 09:42:41.537916 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_baea4578-7ae2-42d3-a07a-ef15c47c886b/ansibletest-ansibletest/0.log" Oct 02 09:42:41 crc kubenswrapper[4960]: I1002 09:42:41.819838 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-69bd96fd88-9gstt_76d6cdce-7551-483b-859c-678805e24895/barbican-api/0.log" Oct 02 09:42:42 crc kubenswrapper[4960]: I1002 09:42:42.059562 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-69bd96fd88-9gstt_76d6cdce-7551-483b-859c-678805e24895/barbican-api-log/0.log" Oct 02 09:42:42 crc kubenswrapper[4960]: I1002 09:42:42.284654 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84cf7698d6-jhmj6_2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7/barbican-keystone-listener/0.log" Oct 02 09:42:42 crc kubenswrapper[4960]: I1002 09:42:42.863688 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bd958db77-xxp6q_cc171097-1fed-426b-9aa6-1c5c905fa5c6/barbican-worker/0.log" Oct 02 09:42:42 crc kubenswrapper[4960]: I1002 09:42:42.985601 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84cf7698d6-jhmj6_2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7/barbican-keystone-listener-log/0.log" Oct 02 09:42:43 crc kubenswrapper[4960]: I1002 09:42:43.127668 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bd958db77-xxp6q_cc171097-1fed-426b-9aa6-1c5c905fa5c6/barbican-worker-log/0.log" Oct 02 09:42:43 crc kubenswrapper[4960]: I1002 09:42:43.405253 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w_9d905943-1348-47e8-81a5-5794bb84f7ff/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:43 crc kubenswrapper[4960]: I1002 09:42:43.649760 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/ceilometer-central-agent/0.log" Oct 02 09:42:43 crc kubenswrapper[4960]: I1002 09:42:43.718754 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/ceilometer-notification-agent/0.log" Oct 02 09:42:43 crc kubenswrapper[4960]: I1002 09:42:43.932856 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/sg-core/0.log" Oct 02 09:42:44 crc kubenswrapper[4960]: I1002 09:42:44.046839 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/proxy-httpd/0.log" Oct 02 09:42:44 crc kubenswrapper[4960]: I1002 09:42:44.286905 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2_363e150c-f1f7-4cd2-9916-e8724ceb0f9a/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:44 crc kubenswrapper[4960]: I1002 09:42:44.335849 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:42:44 crc kubenswrapper[4960]: E1002 09:42:44.337370 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:42:44 crc kubenswrapper[4960]: I1002 09:42:44.488614 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z_05979476-afd0-404e-87ed-23b6706381f2/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:44 crc kubenswrapper[4960]: I1002 09:42:44.710119 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_913ab46c-3d2d-407c-bd43-6f4b9f9e8f58/cinder-api/0.log" Oct 02 09:42:44 crc kubenswrapper[4960]: I1002 09:42:44.755361 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_913ab46c-3d2d-407c-bd43-6f4b9f9e8f58/cinder-api-log/0.log" Oct 02 09:42:45 crc kubenswrapper[4960]: I1002 09:42:45.108278 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0a9c86d8-b033-44fa-9991-08a9aa629b32/probe/0.log" Oct 02 09:42:45 crc kubenswrapper[4960]: I1002 09:42:45.161634 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0a9c86d8-b033-44fa-9991-08a9aa629b32/cinder-backup/0.log" Oct 02 09:42:45 crc kubenswrapper[4960]: I1002 09:42:45.330365 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d/cinder-scheduler/0.log" Oct 02 09:42:45 crc kubenswrapper[4960]: I1002 09:42:45.412250 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d/probe/0.log" Oct 02 09:42:45 crc kubenswrapper[4960]: I1002 09:42:45.596746 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3bebb599-2d7f-4d91-8cf1-eda510fd14f6/cinder-volume/0.log" Oct 02 09:42:45 crc kubenswrapper[4960]: I1002 09:42:45.759422 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3bebb599-2d7f-4d91-8cf1-eda510fd14f6/probe/0.log" Oct 02 09:42:45 crc kubenswrapper[4960]: I1002 09:42:45.832740 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt_947ad109-ebe1-4de0-ad2e-4fa4ae75b238/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:46 crc kubenswrapper[4960]: I1002 09:42:46.031656 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw_7fe7ff43-8e73-408d-bd2d-262e48941ccd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:46 crc kubenswrapper[4960]: I1002 09:42:46.239210 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6785d65565-vfxf8_b169e295-17c4-457d-8832-bb79f85eb5cb/init/0.log" Oct 02 09:42:46 crc kubenswrapper[4960]: I1002 09:42:46.484714 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6785d65565-vfxf8_b169e295-17c4-457d-8832-bb79f85eb5cb/init/0.log" Oct 02 09:42:46 crc kubenswrapper[4960]: I1002 09:42:46.622519 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6785d65565-vfxf8_b169e295-17c4-457d-8832-bb79f85eb5cb/dnsmasq-dns/0.log" Oct 02 09:42:46 crc kubenswrapper[4960]: I1002 09:42:46.929723 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8da35d5f-3f76-4e03-9df7-7d26b23c601d/glance-httpd/0.log" Oct 02 09:42:46 crc kubenswrapper[4960]: I1002 09:42:46.989489 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8da35d5f-3f76-4e03-9df7-7d26b23c601d/glance-log/0.log" Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.193093 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.195324 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7bde8f5a-7f1c-4c7d-a52c-ee1269736978/glance-log/0.log" Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.240794 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7bde8f5a-7f1c-4c7d-a52c-ee1269736978/glance-httpd/0.log" Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.260508 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkw7q"] Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.410264 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gkw7q" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="registry-server" containerID="cri-o://a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf" gracePeriod=2 Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.505679 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-d66c446d8-rxzk8_35082c42-a42f-4eb7-a12c-91c74a6715f4/horizon/0.log" Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.786464 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_179d0cf9-edfa-4b98-8708-1503790d862e/horizontest-tests-horizontest/0.log" Oct 02 09:42:47 crc kubenswrapper[4960]: I1002 09:42:47.965881 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.066640 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-utilities\") pod \"56df1e3a-f004-4228-ab06-3020e74d27b9\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.066815 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7442\" (UniqueName: \"kubernetes.io/projected/56df1e3a-f004-4228-ab06-3020e74d27b9-kube-api-access-r7442\") pod \"56df1e3a-f004-4228-ab06-3020e74d27b9\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.067031 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-catalog-content\") pod \"56df1e3a-f004-4228-ab06-3020e74d27b9\" (UID: \"56df1e3a-f004-4228-ab06-3020e74d27b9\") " Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.080200 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56df1e3a-f004-4228-ab06-3020e74d27b9-kube-api-access-r7442" (OuterVolumeSpecName: "kube-api-access-r7442") pod "56df1e3a-f004-4228-ab06-3020e74d27b9" (UID: "56df1e3a-f004-4228-ab06-3020e74d27b9"). InnerVolumeSpecName "kube-api-access-r7442". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.087503 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-utilities" (OuterVolumeSpecName: "utilities") pod "56df1e3a-f004-4228-ab06-3020e74d27b9" (UID: "56df1e3a-f004-4228-ab06-3020e74d27b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.090390 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56df1e3a-f004-4228-ab06-3020e74d27b9" (UID: "56df1e3a-f004-4228-ab06-3020e74d27b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.134999 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q_5aa32b54-0374-4032-b19f-d1e65557fb95/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.178957 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7442\" (UniqueName: \"kubernetes.io/projected/56df1e3a-f004-4228-ab06-3020e74d27b9-kube-api-access-r7442\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.179052 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.179063 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56df1e3a-f004-4228-ab06-3020e74d27b9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.393131 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6xk6f_e7f8ad39-e87c-435f-b72b-3c59ac00f2b1/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.440251 4960 generic.go:334] "Generic (PLEG): container finished" podID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerID="a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf" exitCode=0 Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.440296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkw7q" event={"ID":"56df1e3a-f004-4228-ab06-3020e74d27b9","Type":"ContainerDied","Data":"a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf"} Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.440326 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkw7q" event={"ID":"56df1e3a-f004-4228-ab06-3020e74d27b9","Type":"ContainerDied","Data":"76cd8d7ec04edfaa1525ff756d0681cab1541e390c0e55124b6148c48790c034"} Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.440343 4960 scope.go:117] "RemoveContainer" containerID="a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.440551 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkw7q" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.474411 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkw7q"] Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.487267 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkw7q"] Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.488327 4960 scope.go:117] "RemoveContainer" containerID="20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.536204 4960 scope.go:117] "RemoveContainer" containerID="a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.586741 4960 scope.go:117] "RemoveContainer" containerID="a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf" Oct 02 09:42:48 crc kubenswrapper[4960]: E1002 09:42:48.587455 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf\": container with ID starting with a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf not found: ID does not exist" containerID="a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.587513 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf"} err="failed to get container status \"a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf\": rpc error: code = NotFound desc = could not find container \"a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf\": container with ID starting with a19462724ef8d695037bcafe89aa366e987588d9b7e1c17a97a1ec44f66914cf not found: ID does not exist" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.587542 4960 scope.go:117] "RemoveContainer" containerID="20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256" Oct 02 09:42:48 crc kubenswrapper[4960]: E1002 09:42:48.587840 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256\": container with ID starting with 20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256 not found: ID does not exist" containerID="20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.587885 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256"} err="failed to get container status \"20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256\": rpc error: code = NotFound desc = could not find container \"20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256\": container with ID starting with 20ce52be0c83741e784aa696a5934e87d1ae9c05a072effcf283f05571401256 not found: ID does not exist" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.587898 4960 scope.go:117] "RemoveContainer" containerID="a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff" Oct 02 09:42:48 crc kubenswrapper[4960]: E1002 09:42:48.588239 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff\": container with ID starting with a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff not found: ID does not exist" containerID="a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.588362 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff"} err="failed to get container status \"a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff\": rpc error: code = NotFound desc = could not find container \"a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff\": container with ID starting with a57b235ebf0b31442695a59f0047d3a3f3c5419a29683d30b145771ac0cffdff not found: ID does not exist" Oct 02 09:42:48 crc kubenswrapper[4960]: I1002 09:42:48.934871 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323201-k4bsb_20148c75-2594-4fa9-b38e-1fdc97757c0f/keystone-cron/0.log" Oct 02 09:42:49 crc kubenswrapper[4960]: I1002 09:42:49.271336 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-d66c446d8-rxzk8_35082c42-a42f-4eb7-a12c-91c74a6715f4/horizon-log/0.log" Oct 02 09:42:49 crc kubenswrapper[4960]: I1002 09:42:49.361270 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323261-sp279_aa245840-cc66-4671-a12c-00295fe7b5a4/keystone-cron/0.log" Oct 02 09:42:49 crc kubenswrapper[4960]: I1002 09:42:49.563301 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_cc53e302-34cd-4f65-9612-bcb47fc6b238/kube-state-metrics/0.log" Oct 02 09:42:49 crc kubenswrapper[4960]: I1002 09:42:49.810694 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc_f2f4ac6d-fe72-4957-bcdb-983b63fa172b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:50 crc kubenswrapper[4960]: I1002 09:42:50.326107 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_5773acec-b7dc-4887-b302-832737da62e9/manila-api/0.log" Oct 02 09:42:50 crc kubenswrapper[4960]: I1002 09:42:50.358520 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" path="/var/lib/kubelet/pods/56df1e3a-f004-4228-ab06-3020e74d27b9/volumes" Oct 02 09:42:50 crc kubenswrapper[4960]: I1002 09:42:50.359626 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_5773acec-b7dc-4887-b302-832737da62e9/manila-api-log/0.log" Oct 02 09:42:50 crc kubenswrapper[4960]: I1002 09:42:50.720891 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_c97772e1-2ae5-425b-8532-be4fbc404d15/manila-scheduler/0.log" Oct 02 09:42:50 crc kubenswrapper[4960]: I1002 09:42:50.794148 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_c97772e1-2ae5-425b-8532-be4fbc404d15/probe/0.log" Oct 02 09:42:50 crc kubenswrapper[4960]: I1002 09:42:50.884498 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-567f974799-pm8bc_ced7f231-e268-4f66-87ae-b32b809bfa6f/keystone-api/0.log" Oct 02 09:42:51 crc kubenswrapper[4960]: I1002 09:42:51.056482 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_29e1c556-f157-4c20-93a7-d0ceecc98439/manila-share/0.log" Oct 02 09:42:51 crc kubenswrapper[4960]: I1002 09:42:51.123178 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_29e1c556-f157-4c20-93a7-d0ceecc98439/probe/0.log" Oct 02 09:42:52 crc kubenswrapper[4960]: I1002 09:42:52.711385 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d8fb8597c-r7rqs_4e13c181-da1b-4da1-9ed7-daa1d05f0f5f/neutron-httpd/0.log" Oct 02 09:42:53 crc kubenswrapper[4960]: I1002 09:42:53.178335 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d8fb8597c-r7rqs_4e13c181-da1b-4da1-9ed7-daa1d05f0f5f/neutron-api/0.log" Oct 02 09:42:53 crc kubenswrapper[4960]: I1002 09:42:53.189329 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw_9708c6b1-7f2f-4820-9072-12f019a1a731/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:42:56 crc kubenswrapper[4960]: I1002 09:42:56.331625 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:42:56 crc kubenswrapper[4960]: E1002 09:42:56.332651 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:42:57 crc kubenswrapper[4960]: I1002 09:42:57.786744 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d889074b-a08e-49cf-8272-2b51fd6a9286/nova-api-log/0.log" Oct 02 09:42:58 crc kubenswrapper[4960]: I1002 09:42:58.810046 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e82f1c13-1224-4c25-9050-d55a2ff7c5e6/nova-cell0-conductor-conductor/0.log" Oct 02 09:42:58 crc kubenswrapper[4960]: I1002 09:42:58.941082 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d889074b-a08e-49cf-8272-2b51fd6a9286/nova-api-api/0.log" Oct 02 09:42:59 crc kubenswrapper[4960]: I1002 09:42:59.343805 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c3bd25c3-7b92-4cc8-a683-3ab41bceccf6/nova-cell1-conductor-conductor/0.log" Oct 02 09:42:59 crc kubenswrapper[4960]: I1002 09:42:59.768752 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_bc95fd9f-83d9-4731-904a-ba493b31656c/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 09:42:59 crc kubenswrapper[4960]: I1002 09:42:59.943545 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z_4a1cec6e-b3b0-48e7-aa5a-8969b06b069f/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:43:00 crc kubenswrapper[4960]: I1002 09:43:00.271601 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a94097fa-b407-419c-9fad-5078a1f8967f/nova-metadata-log/0.log" Oct 02 09:43:01 crc kubenswrapper[4960]: I1002 09:43:01.490808 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_189dd145-b9b7-4dc5-9fae-a44b359f8748/nova-scheduler-scheduler/0.log" Oct 02 09:43:02 crc kubenswrapper[4960]: I1002 09:43:02.024891 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1ea27c7c-a834-41de-9a07-3f0611c3bc23/mysql-bootstrap/0.log" Oct 02 09:43:02 crc kubenswrapper[4960]: I1002 09:43:02.282695 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1ea27c7c-a834-41de-9a07-3f0611c3bc23/mysql-bootstrap/0.log" Oct 02 09:43:02 crc kubenswrapper[4960]: I1002 09:43:02.529654 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1ea27c7c-a834-41de-9a07-3f0611c3bc23/galera/0.log" Oct 02 09:43:03 crc kubenswrapper[4960]: I1002 09:43:03.054511 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3be247d9-d9ed-49f1-9013-2ad50b6e4ed8/mysql-bootstrap/0.log" Oct 02 09:43:03 crc kubenswrapper[4960]: I1002 09:43:03.261031 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3be247d9-d9ed-49f1-9013-2ad50b6e4ed8/mysql-bootstrap/0.log" Oct 02 09:43:03 crc kubenswrapper[4960]: I1002 09:43:03.533959 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3be247d9-d9ed-49f1-9013-2ad50b6e4ed8/galera/0.log" Oct 02 09:43:03 crc kubenswrapper[4960]: I1002 09:43:03.745375 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_34704ecb-b08d-4f67-b6ed-66141a0f6e74/openstackclient/0.log" Oct 02 09:43:04 crc kubenswrapper[4960]: I1002 09:43:04.018395 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-krnzn_ad314560-608d-471a-a6f6-659b9c0755d6/ovn-controller/0.log" Oct 02 09:43:04 crc kubenswrapper[4960]: I1002 09:43:04.206105 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hmxsf_09ac27ea-acdf-444a-b8b8-0f61332416b1/openstack-network-exporter/0.log" Oct 02 09:43:04 crc kubenswrapper[4960]: I1002 09:43:04.668879 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovsdb-server-init/0.log" Oct 02 09:43:04 crc kubenswrapper[4960]: I1002 09:43:04.750169 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovsdb-server-init/0.log" Oct 02 09:43:04 crc kubenswrapper[4960]: I1002 09:43:04.917934 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovs-vswitchd/0.log" Oct 02 09:43:04 crc kubenswrapper[4960]: I1002 09:43:04.951628 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovsdb-server/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.143711 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a94097fa-b407-419c-9fad-5078a1f8967f/nova-metadata-metadata/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.295220 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gdjnm_ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.449404 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_419dd9e4-dbbd-4a58-823b-b987afdf20cb/openstack-network-exporter/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.520067 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_419dd9e4-dbbd-4a58-823b-b987afdf20cb/ovn-northd/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.714025 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_225801bc-ca07-4fa8-8d5c-1d56cfef957b/openstack-network-exporter/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.715020 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_225801bc-ca07-4fa8-8d5c-1d56cfef957b/ovsdbserver-nb/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.935269 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67a98bd9-8fba-44c5-9229-de8a1053d9d6/openstack-network-exporter/0.log" Oct 02 09:43:05 crc kubenswrapper[4960]: I1002 09:43:05.965535 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67a98bd9-8fba-44c5-9229-de8a1053d9d6/ovsdbserver-sb/0.log" Oct 02 09:43:06 crc kubenswrapper[4960]: I1002 09:43:06.729413 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43e28122-07de-46d8-b2bd-918405699fd7/setup-container/0.log" Oct 02 09:43:06 crc kubenswrapper[4960]: I1002 09:43:06.891910 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-64f4f7488b-q47mf_9c89938d-910f-4d87-85a5-22bfc7cc24e9/placement-api/0.log" Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.160894 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43e28122-07de-46d8-b2bd-918405699fd7/setup-container/0.log" Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.244667 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43e28122-07de-46d8-b2bd-918405699fd7/rabbitmq/0.log" Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.299633 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-64f4f7488b-q47mf_9c89938d-910f-4d87-85a5-22bfc7cc24e9/placement-log/0.log" Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.329931 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.516302 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_917895a0-f990-4182-bc76-14bfddefa2a4/setup-container/0.log" Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.641015 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"273166d29cffc5df3a7dd810e2a0ba752cb52673b1bea6051b28005b7941d461"} Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.753664 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_917895a0-f990-4182-bc76-14bfddefa2a4/rabbitmq/0.log" Oct 02 09:43:07 crc kubenswrapper[4960]: I1002 09:43:07.773835 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_917895a0-f990-4182-bc76-14bfddefa2a4/setup-container/0.log" Oct 02 09:43:08 crc kubenswrapper[4960]: I1002 09:43:08.000092 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4_defafea0-a1a3-4e6b-82ce-5b431be9ff5b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:43:08 crc kubenswrapper[4960]: I1002 09:43:08.118221 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w_e621ea76-77e5-4ae4-aee0-eb92a4f247a5/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:43:08 crc kubenswrapper[4960]: I1002 09:43:08.347762 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-phrfr_5066f055-a35b-4ef7-a935-038d15d9838c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:43:08 crc kubenswrapper[4960]: I1002 09:43:08.612222 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-s7hvl_843c7228-62d1-4839-9459-3de9d174f70a/ssh-known-hosts-edpm-deployment/0.log" Oct 02 09:43:08 crc kubenswrapper[4960]: I1002 09:43:08.923574 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_aed3697d-fab0-4c3e-b5d0-21fb0037a412/tempest-tests-tempest-tests-runner/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.055151 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_3cf44a73-42f7-41cc-82e4-f085f154cdbd/tempest-tests-tempest-tests-runner/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.223179 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_a15540db-b8e9-444e-9d55-7bcb2b1c4531/test-operator-logs-container/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.367224 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fded14c9-2fd8-46e6-83a9-84c4ca056120/memcached/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.484576 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f/test-operator-logs-container/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.559046 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_06a6e149-1b05-45a2-98b0-f53b766d60eb/test-operator-logs-container/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.728324 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_6213b3c4-9214-43d0-9ddf-782b3ec673af/test-operator-logs-container/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.838609 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_2b480863-210b-4fb7-8980-e939d5fc9603/tobiko-tests-tobiko/0.log" Oct 02 09:43:09 crc kubenswrapper[4960]: I1002 09:43:09.930535 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8/tobiko-tests-tobiko/0.log" Oct 02 09:43:10 crc kubenswrapper[4960]: I1002 09:43:10.057580 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j_5488228f-8fc0-46cb-be89-51d65f7ccff8/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:43:29 crc kubenswrapper[4960]: E1002 09:43:29.330579 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:44:30 crc kubenswrapper[4960]: I1002 09:44:30.518069 4960 generic.go:334] "Generic (PLEG): container finished" podID="5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c" containerID="1bb393d2c30f15aca66d8d4163d2af2f3f5c21f8c48ba5eba8e60b8a5a382d68" exitCode=0 Oct 02 09:44:30 crc kubenswrapper[4960]: I1002 09:44:30.518200 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" event={"ID":"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c","Type":"ContainerDied","Data":"1bb393d2c30f15aca66d8d4163d2af2f3f5c21f8c48ba5eba8e60b8a5a382d68"} Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.645442 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.686520 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-4h66m"] Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.696862 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-4h66m"] Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.712145 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-host\") pod \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.712378 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-host" (OuterVolumeSpecName: "host") pod "5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c" (UID: "5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.712587 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6t26\" (UniqueName: \"kubernetes.io/projected/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-kube-api-access-z6t26\") pod \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\" (UID: \"5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c\") " Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.713276 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.720997 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-kube-api-access-z6t26" (OuterVolumeSpecName: "kube-api-access-z6t26") pod "5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c" (UID: "5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c"). InnerVolumeSpecName "kube-api-access-z6t26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:31 crc kubenswrapper[4960]: I1002 09:44:31.815953 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6t26\" (UniqueName: \"kubernetes.io/projected/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c-kube-api-access-z6t26\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.343661 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c" path="/var/lib/kubelet/pods/5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c/volumes" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.540883 4960 scope.go:117] "RemoveContainer" containerID="1bb393d2c30f15aca66d8d4163d2af2f3f5c21f8c48ba5eba8e60b8a5a382d68" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.541104 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-4h66m" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.881587 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-6n57h"] Oct 02 09:44:32 crc kubenswrapper[4960]: E1002 09:44:32.882332 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="extract-content" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.882347 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="extract-content" Oct 02 09:44:32 crc kubenswrapper[4960]: E1002 09:44:32.882367 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c" containerName="container-00" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.882374 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c" containerName="container-00" Oct 02 09:44:32 crc kubenswrapper[4960]: E1002 09:44:32.882392 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="registry-server" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.882398 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="registry-server" Oct 02 09:44:32 crc kubenswrapper[4960]: E1002 09:44:32.882409 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="extract-utilities" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.882417 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="extract-utilities" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.882607 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="56df1e3a-f004-4228-ab06-3020e74d27b9" containerName="registry-server" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.882625 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4a6ab5-8335-4d58-9d89-848b7fdbcb5c" containerName="container-00" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.883342 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.942087 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6xjr\" (UniqueName: \"kubernetes.io/projected/efb8db75-87a6-41ea-a9ed-e3d242c54938-kube-api-access-z6xjr\") pod \"crc-debug-6n57h\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:32 crc kubenswrapper[4960]: I1002 09:44:32.942145 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efb8db75-87a6-41ea-a9ed-e3d242c54938-host\") pod \"crc-debug-6n57h\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:33 crc kubenswrapper[4960]: I1002 09:44:33.045167 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6xjr\" (UniqueName: \"kubernetes.io/projected/efb8db75-87a6-41ea-a9ed-e3d242c54938-kube-api-access-z6xjr\") pod \"crc-debug-6n57h\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:33 crc kubenswrapper[4960]: I1002 09:44:33.045223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efb8db75-87a6-41ea-a9ed-e3d242c54938-host\") pod \"crc-debug-6n57h\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:33 crc kubenswrapper[4960]: I1002 09:44:33.045368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efb8db75-87a6-41ea-a9ed-e3d242c54938-host\") pod \"crc-debug-6n57h\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:33 crc kubenswrapper[4960]: I1002 09:44:33.065686 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6xjr\" (UniqueName: \"kubernetes.io/projected/efb8db75-87a6-41ea-a9ed-e3d242c54938-kube-api-access-z6xjr\") pod \"crc-debug-6n57h\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:33 crc kubenswrapper[4960]: I1002 09:44:33.203506 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:33 crc kubenswrapper[4960]: I1002 09:44:33.554134 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" event={"ID":"efb8db75-87a6-41ea-a9ed-e3d242c54938","Type":"ContainerStarted","Data":"f1c7a840a19c1a4ab73165511c6b6fd77372ca2cb8a0ad2ae8170e63aebd2fa0"} Oct 02 09:44:34 crc kubenswrapper[4960]: I1002 09:44:34.565116 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" event={"ID":"efb8db75-87a6-41ea-a9ed-e3d242c54938","Type":"ContainerStarted","Data":"18c80e748dd37b8a7d95d93a5d383d3367a02f6dda48e541291d2948b3e7f3c7"} Oct 02 09:44:34 crc kubenswrapper[4960]: I1002 09:44:34.591467 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" podStartSLOduration=2.5914329069999997 podStartE2EDuration="2.591432907s" podCreationTimestamp="2025-10-02 09:44:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:34.579890231 +0000 UTC m=+8895.611836518" watchObservedRunningTime="2025-10-02 09:44:34.591432907 +0000 UTC m=+8895.623379194" Oct 02 09:44:35 crc kubenswrapper[4960]: E1002 09:44:35.500805 4960 log.go:32] "ReopenContainerLog from runtime service failed" err="rpc error: code = Unknown desc = container is not running" containerID="18c80e748dd37b8a7d95d93a5d383d3367a02f6dda48e541291d2948b3e7f3c7" Oct 02 09:44:35 crc kubenswrapper[4960]: E1002 09:44:35.500933 4960 container_log_manager.go:307] "Failed to rotate log for container" err="failed to rotate log \"/var/log/pods/openshift-must-gather-2k8qc_crc-debug-6n57h_efb8db75-87a6-41ea-a9ed-e3d242c54938/container-00/0.log\": failed to reopen container log \"18c80e748dd37b8a7d95d93a5d383d3367a02f6dda48e541291d2948b3e7f3c7\": rpc error: code = Unknown desc = container is not running" worker=1 containerID="18c80e748dd37b8a7d95d93a5d383d3367a02f6dda48e541291d2948b3e7f3c7" path="/var/log/pods/openshift-must-gather-2k8qc_crc-debug-6n57h_efb8db75-87a6-41ea-a9ed-e3d242c54938/container-00/0.log" currentSize=81634817 maxSize=52428800 Oct 02 09:44:36 crc kubenswrapper[4960]: I1002 09:44:36.585625 4960 generic.go:334] "Generic (PLEG): container finished" podID="efb8db75-87a6-41ea-a9ed-e3d242c54938" containerID="18c80e748dd37b8a7d95d93a5d383d3367a02f6dda48e541291d2948b3e7f3c7" exitCode=0 Oct 02 09:44:36 crc kubenswrapper[4960]: I1002 09:44:36.585693 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" event={"ID":"efb8db75-87a6-41ea-a9ed-e3d242c54938","Type":"ContainerDied","Data":"18c80e748dd37b8a7d95d93a5d383d3367a02f6dda48e541291d2948b3e7f3c7"} Oct 02 09:44:37 crc kubenswrapper[4960]: I1002 09:44:37.711811 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:37 crc kubenswrapper[4960]: I1002 09:44:37.738132 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efb8db75-87a6-41ea-a9ed-e3d242c54938-host\") pod \"efb8db75-87a6-41ea-a9ed-e3d242c54938\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " Oct 02 09:44:37 crc kubenswrapper[4960]: I1002 09:44:37.738269 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efb8db75-87a6-41ea-a9ed-e3d242c54938-host" (OuterVolumeSpecName: "host") pod "efb8db75-87a6-41ea-a9ed-e3d242c54938" (UID: "efb8db75-87a6-41ea-a9ed-e3d242c54938"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:37 crc kubenswrapper[4960]: I1002 09:44:37.738895 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6xjr\" (UniqueName: \"kubernetes.io/projected/efb8db75-87a6-41ea-a9ed-e3d242c54938-kube-api-access-z6xjr\") pod \"efb8db75-87a6-41ea-a9ed-e3d242c54938\" (UID: \"efb8db75-87a6-41ea-a9ed-e3d242c54938\") " Oct 02 09:44:37 crc kubenswrapper[4960]: I1002 09:44:37.739891 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efb8db75-87a6-41ea-a9ed-e3d242c54938-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:37 crc kubenswrapper[4960]: I1002 09:44:37.749556 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb8db75-87a6-41ea-a9ed-e3d242c54938-kube-api-access-z6xjr" (OuterVolumeSpecName: "kube-api-access-z6xjr") pod "efb8db75-87a6-41ea-a9ed-e3d242c54938" (UID: "efb8db75-87a6-41ea-a9ed-e3d242c54938"). InnerVolumeSpecName "kube-api-access-z6xjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:37 crc kubenswrapper[4960]: I1002 09:44:37.841658 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6xjr\" (UniqueName: \"kubernetes.io/projected/efb8db75-87a6-41ea-a9ed-e3d242c54938-kube-api-access-z6xjr\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:38 crc kubenswrapper[4960]: I1002 09:44:38.614468 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" event={"ID":"efb8db75-87a6-41ea-a9ed-e3d242c54938","Type":"ContainerDied","Data":"f1c7a840a19c1a4ab73165511c6b6fd77372ca2cb8a0ad2ae8170e63aebd2fa0"} Oct 02 09:44:38 crc kubenswrapper[4960]: I1002 09:44:38.614525 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1c7a840a19c1a4ab73165511c6b6fd77372ca2cb8a0ad2ae8170e63aebd2fa0" Oct 02 09:44:38 crc kubenswrapper[4960]: I1002 09:44:38.614551 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-6n57h" Oct 02 09:44:46 crc kubenswrapper[4960]: I1002 09:44:46.591202 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-6n57h"] Oct 02 09:44:46 crc kubenswrapper[4960]: I1002 09:44:46.600224 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-6n57h"] Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.784404 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-bm69f"] Oct 02 09:44:47 crc kubenswrapper[4960]: E1002 09:44:47.784951 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb8db75-87a6-41ea-a9ed-e3d242c54938" containerName="container-00" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.784966 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb8db75-87a6-41ea-a9ed-e3d242c54938" containerName="container-00" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.785246 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb8db75-87a6-41ea-a9ed-e3d242c54938" containerName="container-00" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.786518 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.888403 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvfbz\" (UniqueName: \"kubernetes.io/projected/387aeba4-a3ee-4c5d-9158-551a8f8d469d-kube-api-access-nvfbz\") pod \"crc-debug-bm69f\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.888561 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/387aeba4-a3ee-4c5d-9158-551a8f8d469d-host\") pod \"crc-debug-bm69f\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.991030 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/387aeba4-a3ee-4c5d-9158-551a8f8d469d-host\") pod \"crc-debug-bm69f\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.991241 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/387aeba4-a3ee-4c5d-9158-551a8f8d469d-host\") pod \"crc-debug-bm69f\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:47 crc kubenswrapper[4960]: I1002 09:44:47.991261 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvfbz\" (UniqueName: \"kubernetes.io/projected/387aeba4-a3ee-4c5d-9158-551a8f8d469d-kube-api-access-nvfbz\") pod \"crc-debug-bm69f\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.011837 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvfbz\" (UniqueName: \"kubernetes.io/projected/387aeba4-a3ee-4c5d-9158-551a8f8d469d-kube-api-access-nvfbz\") pod \"crc-debug-bm69f\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.116033 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.346160 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efb8db75-87a6-41ea-a9ed-e3d242c54938" path="/var/lib/kubelet/pods/efb8db75-87a6-41ea-a9ed-e3d242c54938/volumes" Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.851492 4960 generic.go:334] "Generic (PLEG): container finished" podID="387aeba4-a3ee-4c5d-9158-551a8f8d469d" containerID="897cda308866c2aa89b035c48f4662322a48e0b50010b90f3181556af4b0e100" exitCode=0 Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.851555 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-bm69f" event={"ID":"387aeba4-a3ee-4c5d-9158-551a8f8d469d","Type":"ContainerDied","Data":"897cda308866c2aa89b035c48f4662322a48e0b50010b90f3181556af4b0e100"} Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.851595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/crc-debug-bm69f" event={"ID":"387aeba4-a3ee-4c5d-9158-551a8f8d469d","Type":"ContainerStarted","Data":"1f18114093197f004e55d7775bcdb5df11e33d94527cb5de0edd6dc6b16e52c1"} Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.900111 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-bm69f"] Oct 02 09:44:48 crc kubenswrapper[4960]: I1002 09:44:48.914850 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2k8qc/crc-debug-bm69f"] Oct 02 09:44:49 crc kubenswrapper[4960]: I1002 09:44:49.983548 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.045871 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvfbz\" (UniqueName: \"kubernetes.io/projected/387aeba4-a3ee-4c5d-9158-551a8f8d469d-kube-api-access-nvfbz\") pod \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.046129 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/387aeba4-a3ee-4c5d-9158-551a8f8d469d-host\") pod \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\" (UID: \"387aeba4-a3ee-4c5d-9158-551a8f8d469d\") " Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.046293 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/387aeba4-a3ee-4c5d-9158-551a8f8d469d-host" (OuterVolumeSpecName: "host") pod "387aeba4-a3ee-4c5d-9158-551a8f8d469d" (UID: "387aeba4-a3ee-4c5d-9158-551a8f8d469d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.047051 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/387aeba4-a3ee-4c5d-9158-551a8f8d469d-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.062460 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/387aeba4-a3ee-4c5d-9158-551a8f8d469d-kube-api-access-nvfbz" (OuterVolumeSpecName: "kube-api-access-nvfbz") pod "387aeba4-a3ee-4c5d-9158-551a8f8d469d" (UID: "387aeba4-a3ee-4c5d-9158-551a8f8d469d"). InnerVolumeSpecName "kube-api-access-nvfbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.148988 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvfbz\" (UniqueName: \"kubernetes.io/projected/387aeba4-a3ee-4c5d-9158-551a8f8d469d-kube-api-access-nvfbz\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.343763 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="387aeba4-a3ee-4c5d-9158-551a8f8d469d" path="/var/lib/kubelet/pods/387aeba4-a3ee-4c5d-9158-551a8f8d469d/volumes" Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.876042 4960 scope.go:117] "RemoveContainer" containerID="897cda308866c2aa89b035c48f4662322a48e0b50010b90f3181556af4b0e100" Oct 02 09:44:50 crc kubenswrapper[4960]: I1002 09:44:50.876192 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/crc-debug-bm69f" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.088621 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/util/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.304205 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/pull/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.312868 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/pull/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.388197 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/util/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.563314 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/extract/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.593371 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/util/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.594798 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/pull/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.773820 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-mfg89_b6b8a229-03cf-4cb0-ab8a-a3133dadfc21/kube-rbac-proxy/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.866246 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-mfg89_b6b8a229-03cf-4cb0-ab8a-a3133dadfc21/manager/0.log" Oct 02 09:44:51 crc kubenswrapper[4960]: I1002 09:44:51.931929 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-dfpnm_539fe2a2-8989-4fb6-a24e-dee67304e96b/kube-rbac-proxy/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.055707 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-dfpnm_539fe2a2-8989-4fb6-a24e-dee67304e96b/manager/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.144922 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-7qjq7_6c71ae76-8a1e-4570-8950-8cf4962169e4/kube-rbac-proxy/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.234210 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-7qjq7_6c71ae76-8a1e-4570-8950-8cf4962169e4/manager/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.354821 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-5mk92_b9dd3b02-9840-4eef-b378-b7888406a57c/kube-rbac-proxy/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.450477 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-5mk92_b9dd3b02-9840-4eef-b378-b7888406a57c/manager/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.582217 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-g9bwh_32ba83c2-5436-4b71-9a95-e552613c6bff/kube-rbac-proxy/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.641674 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-g9bwh_32ba83c2-5436-4b71-9a95-e552613c6bff/manager/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.803328 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-cd9cz_72b0c3f8-2a66-42da-aa83-298bda8f332d/kube-rbac-proxy/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.860133 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-cd9cz_72b0c3f8-2a66-42da-aa83-298bda8f332d/manager/0.log" Oct 02 09:44:52 crc kubenswrapper[4960]: I1002 09:44:52.950074 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-hbzp7_cf515440-a75b-4f32-8c08-2f108da16b56/kube-rbac-proxy/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.174433 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-wm4wc_4bcfb151-7975-4567-9a1f-0cb7dd128696/kube-rbac-proxy/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.207117 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-wm4wc_4bcfb151-7975-4567-9a1f-0cb7dd128696/manager/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.231326 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-hbzp7_cf515440-a75b-4f32-8c08-2f108da16b56/manager/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.465176 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-54crk_8c352774-9d31-4d0a-a1ee-0a96bfbf6239/kube-rbac-proxy/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.516959 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-54crk_8c352774-9d31-4d0a-a1ee-0a96bfbf6239/manager/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.634356 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-hkswg_2af8b0c5-6356-4c12-894e-1bb39e4bf981/kube-rbac-proxy/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.783243 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-kgzd2_5d7077dd-5b52-49af-9dd7-c81f361139ec/kube-rbac-proxy/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.791305 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-hkswg_2af8b0c5-6356-4c12-894e-1bb39e4bf981/manager/0.log" Oct 02 09:44:53 crc kubenswrapper[4960]: I1002 09:44:53.922402 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-kgzd2_5d7077dd-5b52-49af-9dd7-c81f361139ec/manager/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.046296 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-6lpjs_c0d6bd86-ab92-4f37-be9c-003aab10f910/kube-rbac-proxy/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.295033 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-6lpjs_c0d6bd86-ab92-4f37-be9c-003aab10f910/manager/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.457376 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-4vtmz_87d511d4-0cbf-49b2-a2e2-b2e2fc726f60/kube-rbac-proxy/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.522693 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xhd8b_418a02f4-eb9e-4079-969a-c75f8307b078/kube-rbac-proxy/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.587267 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-4vtmz_87d511d4-0cbf-49b2-a2e2-b2e2fc726f60/manager/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.612511 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xhd8b_418a02f4-eb9e-4079-969a-c75f8307b078/manager/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.767799 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-787874f5b776jcd_70d6affd-3067-440a-8b1e-69285828be97/kube-rbac-proxy/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.807920 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-787874f5b776jcd_70d6affd-3067-440a-8b1e-69285828be97/manager/0.log" Oct 02 09:44:54 crc kubenswrapper[4960]: I1002 09:44:54.849394 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-67698bcd47-nz8r6_f802e6c8-39a5-4acf-b85c-29f235283937/kube-rbac-proxy/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.072188 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-57f4f67bc5-p8hvz_cefbfffd-77e3-4fef-a4b6-b4f94a7cee80/kube-rbac-proxy/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.131347 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-57f4f67bc5-p8hvz_cefbfffd-77e3-4fef-a4b6-b4f94a7cee80/operator/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.327616 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-lk5kd_a786a744-833c-40f9-91a7-8b5f312a626c/kube-rbac-proxy/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: E1002 09:44:55.329572 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.496562 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-lk5kd_a786a744-833c-40f9-91a7-8b5f312a626c/manager/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.627195 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-htqft_12f72b6d-7325-428c-8a30-fd0d75158b31/registry-server/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.717275 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-fkgnk_de18430c-640c-46ac-9eb5-de6216332a80/kube-rbac-proxy/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.830746 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-fkgnk_de18430c-640c-46ac-9eb5-de6216332a80/manager/0.log" Oct 02 09:44:55 crc kubenswrapper[4960]: I1002 09:44:55.919633 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk_2fac2998-8d33-448f-877a-388b6570a141/operator/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.108868 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-lzlkk_e52d2a9e-b404-44cf-978a-a2fed39d6337/kube-rbac-proxy/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.111442 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-lzlkk_e52d2a9e-b404-44cf-978a-a2fed39d6337/manager/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.263240 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-cmrs5_73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894/kube-rbac-proxy/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.534331 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-fd7b549cd-82rf9_868e4045-9e39-4add-9023-4ef41caed499/kube-rbac-proxy/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.538719 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-67698bcd47-nz8r6_f802e6c8-39a5-4acf-b85c-29f235283937/manager/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.558772 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-cmrs5_73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894/manager/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.599146 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-fd7b549cd-82rf9_868e4045-9e39-4add-9023-4ef41caed499/manager/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.762471 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-69k7g_97605d31-7ce2-4f75-848a-1fcf39e54d3e/kube-rbac-proxy/0.log" Oct 02 09:44:56 crc kubenswrapper[4960]: I1002 09:44:56.795658 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-69k7g_97605d31-7ce2-4f75-848a-1fcf39e54d3e/manager/0.log" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.183667 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b"] Oct 02 09:45:00 crc kubenswrapper[4960]: E1002 09:45:00.184861 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387aeba4-a3ee-4c5d-9158-551a8f8d469d" containerName="container-00" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.184875 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="387aeba4-a3ee-4c5d-9158-551a8f8d469d" containerName="container-00" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.185085 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="387aeba4-a3ee-4c5d-9158-551a8f8d469d" containerName="container-00" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.185735 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.188221 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.188622 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.213251 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b"] Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.282724 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50203768-d1a0-475d-b543-e814fab06945-secret-volume\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.282820 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km6rh\" (UniqueName: \"kubernetes.io/projected/50203768-d1a0-475d-b543-e814fab06945-kube-api-access-km6rh\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.283020 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50203768-d1a0-475d-b543-e814fab06945-config-volume\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.392806 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50203768-d1a0-475d-b543-e814fab06945-config-volume\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.392952 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50203768-d1a0-475d-b543-e814fab06945-secret-volume\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.393029 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km6rh\" (UniqueName: \"kubernetes.io/projected/50203768-d1a0-475d-b543-e814fab06945-kube-api-access-km6rh\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.396632 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50203768-d1a0-475d-b543-e814fab06945-config-volume\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.406431 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50203768-d1a0-475d-b543-e814fab06945-secret-volume\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.426142 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km6rh\" (UniqueName: \"kubernetes.io/projected/50203768-d1a0-475d-b543-e814fab06945-kube-api-access-km6rh\") pod \"collect-profiles-29323305-2fw8b\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:00 crc kubenswrapper[4960]: I1002 09:45:00.518727 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:01 crc kubenswrapper[4960]: I1002 09:45:01.147889 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b"] Oct 02 09:45:02 crc kubenswrapper[4960]: I1002 09:45:02.010254 4960 generic.go:334] "Generic (PLEG): container finished" podID="50203768-d1a0-475d-b543-e814fab06945" containerID="e94c5dd35193789a785027268ae0d0f03eec8941e1baf18bac3b4951c17d2467" exitCode=0 Oct 02 09:45:02 crc kubenswrapper[4960]: I1002 09:45:02.010378 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" event={"ID":"50203768-d1a0-475d-b543-e814fab06945","Type":"ContainerDied","Data":"e94c5dd35193789a785027268ae0d0f03eec8941e1baf18bac3b4951c17d2467"} Oct 02 09:45:02 crc kubenswrapper[4960]: I1002 09:45:02.010758 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" event={"ID":"50203768-d1a0-475d-b543-e814fab06945","Type":"ContainerStarted","Data":"338c34acc0f103fad747fc821d450f6f09874e79bc73a726b880c6e035298240"} Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.465450 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.568091 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50203768-d1a0-475d-b543-e814fab06945-secret-volume\") pod \"50203768-d1a0-475d-b543-e814fab06945\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.568271 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50203768-d1a0-475d-b543-e814fab06945-config-volume\") pod \"50203768-d1a0-475d-b543-e814fab06945\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.568449 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km6rh\" (UniqueName: \"kubernetes.io/projected/50203768-d1a0-475d-b543-e814fab06945-kube-api-access-km6rh\") pod \"50203768-d1a0-475d-b543-e814fab06945\" (UID: \"50203768-d1a0-475d-b543-e814fab06945\") " Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.568937 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50203768-d1a0-475d-b543-e814fab06945-config-volume" (OuterVolumeSpecName: "config-volume") pod "50203768-d1a0-475d-b543-e814fab06945" (UID: "50203768-d1a0-475d-b543-e814fab06945"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.574483 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50203768-d1a0-475d-b543-e814fab06945-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "50203768-d1a0-475d-b543-e814fab06945" (UID: "50203768-d1a0-475d-b543-e814fab06945"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.584080 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50203768-d1a0-475d-b543-e814fab06945-kube-api-access-km6rh" (OuterVolumeSpecName: "kube-api-access-km6rh") pod "50203768-d1a0-475d-b543-e814fab06945" (UID: "50203768-d1a0-475d-b543-e814fab06945"). InnerVolumeSpecName "kube-api-access-km6rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.671313 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km6rh\" (UniqueName: \"kubernetes.io/projected/50203768-d1a0-475d-b543-e814fab06945-kube-api-access-km6rh\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.671353 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50203768-d1a0-475d-b543-e814fab06945-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:03 crc kubenswrapper[4960]: I1002 09:45:03.671365 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50203768-d1a0-475d-b543-e814fab06945-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:04 crc kubenswrapper[4960]: I1002 09:45:04.033267 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" event={"ID":"50203768-d1a0-475d-b543-e814fab06945","Type":"ContainerDied","Data":"338c34acc0f103fad747fc821d450f6f09874e79bc73a726b880c6e035298240"} Oct 02 09:45:04 crc kubenswrapper[4960]: I1002 09:45:04.033635 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="338c34acc0f103fad747fc821d450f6f09874e79bc73a726b880c6e035298240" Oct 02 09:45:04 crc kubenswrapper[4960]: I1002 09:45:04.033384 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-2fw8b" Oct 02 09:45:04 crc kubenswrapper[4960]: I1002 09:45:04.535712 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846"] Oct 02 09:45:04 crc kubenswrapper[4960]: I1002 09:45:04.549530 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-zl846"] Oct 02 09:45:06 crc kubenswrapper[4960]: I1002 09:45:06.341227 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7" path="/var/lib/kubelet/pods/8abfb90f-27ca-4fb0-8bed-2c4a76ab38a7/volumes" Oct 02 09:45:15 crc kubenswrapper[4960]: I1002 09:45:15.174252 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4xjvf_3d80dacd-d5ee-4616-878d-0563d69b6b6f/control-plane-machine-set-operator/0.log" Oct 02 09:45:15 crc kubenswrapper[4960]: I1002 09:45:15.400965 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zcpbw_4e0d9ffd-9187-4446-9a5b-87b1b999ff6e/machine-api-operator/0.log" Oct 02 09:45:15 crc kubenswrapper[4960]: I1002 09:45:15.421964 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zcpbw_4e0d9ffd-9187-4446-9a5b-87b1b999ff6e/kube-rbac-proxy/0.log" Oct 02 09:45:26 crc kubenswrapper[4960]: I1002 09:45:26.836941 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vn5rf_aa123932-baa6-435e-b311-e6a62417f39f/cert-manager-controller/0.log" Oct 02 09:45:27 crc kubenswrapper[4960]: I1002 09:45:27.042479 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-74597_c9ca51b7-0180-491c-9d6d-730bea696f09/cert-manager-cainjector/0.log" Oct 02 09:45:27 crc kubenswrapper[4960]: I1002 09:45:27.107990 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-zh4pv_3a34826f-fbd3-4abe-a67f-adc680605fb9/cert-manager-webhook/0.log" Oct 02 09:45:29 crc kubenswrapper[4960]: I1002 09:45:29.150228 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:45:29 crc kubenswrapper[4960]: I1002 09:45:29.151502 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.323345 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mb5p9"] Oct 02 09:45:31 crc kubenswrapper[4960]: E1002 09:45:31.324333 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50203768-d1a0-475d-b543-e814fab06945" containerName="collect-profiles" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.324354 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="50203768-d1a0-475d-b543-e814fab06945" containerName="collect-profiles" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.324615 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="50203768-d1a0-475d-b543-e814fab06945" containerName="collect-profiles" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.326516 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.339232 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mb5p9"] Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.399498 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-catalog-content\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.399808 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm9r2\" (UniqueName: \"kubernetes.io/projected/239f1025-ea2f-4d30-81b5-d964fe633cd1-kube-api-access-gm9r2\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.399921 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-utilities\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.502020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-catalog-content\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.502087 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm9r2\" (UniqueName: \"kubernetes.io/projected/239f1025-ea2f-4d30-81b5-d964fe633cd1-kube-api-access-gm9r2\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.502139 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-utilities\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.502658 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-catalog-content\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.502700 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-utilities\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.527015 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm9r2\" (UniqueName: \"kubernetes.io/projected/239f1025-ea2f-4d30-81b5-d964fe633cd1-kube-api-access-gm9r2\") pod \"certified-operators-mb5p9\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:31 crc kubenswrapper[4960]: I1002 09:45:31.653136 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:32 crc kubenswrapper[4960]: I1002 09:45:32.257937 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mb5p9"] Oct 02 09:45:32 crc kubenswrapper[4960]: I1002 09:45:32.349754 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb5p9" event={"ID":"239f1025-ea2f-4d30-81b5-d964fe633cd1","Type":"ContainerStarted","Data":"1c20842e9613f68d494611575caddea00cd5a52d663af3c9f68d59d5d9cb1ffc"} Oct 02 09:45:33 crc kubenswrapper[4960]: I1002 09:45:33.357810 4960 generic.go:334] "Generic (PLEG): container finished" podID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerID="21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13" exitCode=0 Oct 02 09:45:33 crc kubenswrapper[4960]: I1002 09:45:33.357867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb5p9" event={"ID":"239f1025-ea2f-4d30-81b5-d964fe633cd1","Type":"ContainerDied","Data":"21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13"} Oct 02 09:45:34 crc kubenswrapper[4960]: I1002 09:45:34.368611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb5p9" event={"ID":"239f1025-ea2f-4d30-81b5-d964fe633cd1","Type":"ContainerStarted","Data":"beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be"} Oct 02 09:45:36 crc kubenswrapper[4960]: I1002 09:45:36.390486 4960 generic.go:334] "Generic (PLEG): container finished" podID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerID="beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be" exitCode=0 Oct 02 09:45:36 crc kubenswrapper[4960]: I1002 09:45:36.390558 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb5p9" event={"ID":"239f1025-ea2f-4d30-81b5-d964fe633cd1","Type":"ContainerDied","Data":"beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be"} Oct 02 09:45:37 crc kubenswrapper[4960]: I1002 09:45:37.402357 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb5p9" event={"ID":"239f1025-ea2f-4d30-81b5-d964fe633cd1","Type":"ContainerStarted","Data":"bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285"} Oct 02 09:45:37 crc kubenswrapper[4960]: I1002 09:45:37.429523 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mb5p9" podStartSLOduration=2.794172789 podStartE2EDuration="6.429504198s" podCreationTimestamp="2025-10-02 09:45:31 +0000 UTC" firstStartedPulling="2025-10-02 09:45:33.359964855 +0000 UTC m=+8954.391911142" lastFinishedPulling="2025-10-02 09:45:36.995296264 +0000 UTC m=+8958.027242551" observedRunningTime="2025-10-02 09:45:37.424440033 +0000 UTC m=+8958.456386320" watchObservedRunningTime="2025-10-02 09:45:37.429504198 +0000 UTC m=+8958.461450475" Oct 02 09:45:40 crc kubenswrapper[4960]: I1002 09:45:40.028491 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-k8xfq_d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb/nmstate-console-plugin/0.log" Oct 02 09:45:40 crc kubenswrapper[4960]: I1002 09:45:40.277492 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-77kzr_3242a27c-0911-439a-b346-0f921d212cb8/kube-rbac-proxy/0.log" Oct 02 09:45:40 crc kubenswrapper[4960]: I1002 09:45:40.288601 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-r24b6_bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67/nmstate-handler/0.log" Oct 02 09:45:40 crc kubenswrapper[4960]: I1002 09:45:40.309906 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-77kzr_3242a27c-0911-439a-b346-0f921d212cb8/nmstate-metrics/0.log" Oct 02 09:45:40 crc kubenswrapper[4960]: I1002 09:45:40.537957 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-vpq7s_e1c55993-cf88-4aac-a665-92e26a69a841/nmstate-operator/0.log" Oct 02 09:45:40 crc kubenswrapper[4960]: I1002 09:45:40.591198 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-cjznw_b1cd06a6-a984-40d3-bd87-e81eedce2c7c/nmstate-webhook/0.log" Oct 02 09:45:41 crc kubenswrapper[4960]: I1002 09:45:41.653571 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:41 crc kubenswrapper[4960]: I1002 09:45:41.653732 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:41 crc kubenswrapper[4960]: I1002 09:45:41.710226 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:42 crc kubenswrapper[4960]: I1002 09:45:42.508794 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:42 crc kubenswrapper[4960]: I1002 09:45:42.569407 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mb5p9"] Oct 02 09:45:44 crc kubenswrapper[4960]: I1002 09:45:44.476507 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mb5p9" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="registry-server" containerID="cri-o://bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285" gracePeriod=2 Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.039815 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.124777 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-utilities\") pod \"239f1025-ea2f-4d30-81b5-d964fe633cd1\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.124963 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm9r2\" (UniqueName: \"kubernetes.io/projected/239f1025-ea2f-4d30-81b5-d964fe633cd1-kube-api-access-gm9r2\") pod \"239f1025-ea2f-4d30-81b5-d964fe633cd1\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.125141 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-catalog-content\") pod \"239f1025-ea2f-4d30-81b5-d964fe633cd1\" (UID: \"239f1025-ea2f-4d30-81b5-d964fe633cd1\") " Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.125941 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-utilities" (OuterVolumeSpecName: "utilities") pod "239f1025-ea2f-4d30-81b5-d964fe633cd1" (UID: "239f1025-ea2f-4d30-81b5-d964fe633cd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.138341 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239f1025-ea2f-4d30-81b5-d964fe633cd1-kube-api-access-gm9r2" (OuterVolumeSpecName: "kube-api-access-gm9r2") pod "239f1025-ea2f-4d30-81b5-d964fe633cd1" (UID: "239f1025-ea2f-4d30-81b5-d964fe633cd1"). InnerVolumeSpecName "kube-api-access-gm9r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.182452 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "239f1025-ea2f-4d30-81b5-d964fe633cd1" (UID: "239f1025-ea2f-4d30-81b5-d964fe633cd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.227543 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm9r2\" (UniqueName: \"kubernetes.io/projected/239f1025-ea2f-4d30-81b5-d964fe633cd1-kube-api-access-gm9r2\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.227579 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.227588 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239f1025-ea2f-4d30-81b5-d964fe633cd1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.489129 4960 generic.go:334] "Generic (PLEG): container finished" podID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerID="bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285" exitCode=0 Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.489189 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mb5p9" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.489190 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb5p9" event={"ID":"239f1025-ea2f-4d30-81b5-d964fe633cd1","Type":"ContainerDied","Data":"bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285"} Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.489336 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb5p9" event={"ID":"239f1025-ea2f-4d30-81b5-d964fe633cd1","Type":"ContainerDied","Data":"1c20842e9613f68d494611575caddea00cd5a52d663af3c9f68d59d5d9cb1ffc"} Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.489373 4960 scope.go:117] "RemoveContainer" containerID="bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.516991 4960 scope.go:117] "RemoveContainer" containerID="beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.540256 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mb5p9"] Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.551085 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mb5p9"] Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.560246 4960 scope.go:117] "RemoveContainer" containerID="21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.614054 4960 scope.go:117] "RemoveContainer" containerID="bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285" Oct 02 09:45:45 crc kubenswrapper[4960]: E1002 09:45:45.614661 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285\": container with ID starting with bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285 not found: ID does not exist" containerID="bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.614818 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285"} err="failed to get container status \"bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285\": rpc error: code = NotFound desc = could not find container \"bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285\": container with ID starting with bfaaf04a84688dec065077713a8243d52a380e6a50f0fba212bc7b5ea6084285 not found: ID does not exist" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.614912 4960 scope.go:117] "RemoveContainer" containerID="beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be" Oct 02 09:45:45 crc kubenswrapper[4960]: E1002 09:45:45.615263 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be\": container with ID starting with beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be not found: ID does not exist" containerID="beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.615372 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be"} err="failed to get container status \"beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be\": rpc error: code = NotFound desc = could not find container \"beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be\": container with ID starting with beb6c4086f655700b892ac1156151283ccece2561c7b095b87f6ba1507de83be not found: ID does not exist" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.615495 4960 scope.go:117] "RemoveContainer" containerID="21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13" Oct 02 09:45:45 crc kubenswrapper[4960]: E1002 09:45:45.615823 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13\": container with ID starting with 21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13 not found: ID does not exist" containerID="21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13" Oct 02 09:45:45 crc kubenswrapper[4960]: I1002 09:45:45.615927 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13"} err="failed to get container status \"21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13\": rpc error: code = NotFound desc = could not find container \"21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13\": container with ID starting with 21e926ffa02e90ba08160254d345942d1a3c7f53c3022ac9f216e11b86f73c13 not found: ID does not exist" Oct 02 09:45:46 crc kubenswrapper[4960]: I1002 09:45:46.341452 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" path="/var/lib/kubelet/pods/239f1025-ea2f-4d30-81b5-d964fe633cd1/volumes" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.292528 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rd2k2_2d48cc2b-907f-4811-8cb0-779e2799638d/kube-rbac-proxy/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.407729 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rd2k2_2d48cc2b-907f-4811-8cb0-779e2799638d/controller/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.505924 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.709812 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.746811 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.772946 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.783093 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.936934 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.945136 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.976251 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:45:54 crc kubenswrapper[4960]: I1002 09:45:54.981391 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.175396 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.176101 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.185679 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/controller/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.197021 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.365623 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/frr-metrics/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.383715 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/kube-rbac-proxy-frr/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.384732 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/kube-rbac-proxy/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.592869 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/reloader/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.608062 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-68xcx_ab470676-e7e9-40a7-96db-3b1cc494dc0a/frr-k8s-webhook-server/0.log" Oct 02 09:45:55 crc kubenswrapper[4960]: I1002 09:45:55.874093 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-cf55b8d8b-rzqpg_b96bb98b-d48b-47d6-8357-0c69cd9052c1/manager/0.log" Oct 02 09:45:56 crc kubenswrapper[4960]: I1002 09:45:56.119793 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5j29v_a4cae542-7390-4af8-8c12-97addccc5b4b/kube-rbac-proxy/0.log" Oct 02 09:45:56 crc kubenswrapper[4960]: I1002 09:45:56.169544 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-db8949b7d-zx4b9_8d221df4-ed14-4f8d-82db-6326eaa53a3d/webhook-server/0.log" Oct 02 09:45:57 crc kubenswrapper[4960]: I1002 09:45:57.505037 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5j29v_a4cae542-7390-4af8-8c12-97addccc5b4b/speaker/0.log" Oct 02 09:45:58 crc kubenswrapper[4960]: I1002 09:45:58.433796 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/frr/0.log" Oct 02 09:45:59 crc kubenswrapper[4960]: I1002 09:45:59.150813 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:45:59 crc kubenswrapper[4960]: I1002 09:45:59.150898 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:46:05 crc kubenswrapper[4960]: I1002 09:46:05.441479 4960 scope.go:117] "RemoveContainer" containerID="fa1148cccb5afb54b95d40792734f039630bf557bf19a7c2068b9be73fc92411" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.247249 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/util/0.log" Oct 02 09:46:10 crc kubenswrapper[4960]: E1002 09:46:10.346262 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.419256 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/util/0.log" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.457084 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/pull/0.log" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.495026 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/pull/0.log" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.695948 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/extract/0.log" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.700025 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/util/0.log" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.784668 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/pull/0.log" Oct 02 09:46:10 crc kubenswrapper[4960]: I1002 09:46:10.877060 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-utilities/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.085730 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-utilities/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.088719 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-content/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.118408 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-content/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.358393 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-content/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.374883 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-utilities/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.596859 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-utilities/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.764593 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-utilities/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.878188 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-content/0.log" Oct 02 09:46:11 crc kubenswrapper[4960]: I1002 09:46:11.893472 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-content/0.log" Oct 02 09:46:12 crc kubenswrapper[4960]: I1002 09:46:12.090240 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-utilities/0.log" Oct 02 09:46:12 crc kubenswrapper[4960]: I1002 09:46:12.128988 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-content/0.log" Oct 02 09:46:12 crc kubenswrapper[4960]: I1002 09:46:12.397469 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/util/0.log" Oct 02 09:46:12 crc kubenswrapper[4960]: I1002 09:46:12.592053 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/util/0.log" Oct 02 09:46:12 crc kubenswrapper[4960]: I1002 09:46:12.702168 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/pull/0.log" Oct 02 09:46:12 crc kubenswrapper[4960]: I1002 09:46:12.853238 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/pull/0.log" Oct 02 09:46:13 crc kubenswrapper[4960]: I1002 09:46:13.067296 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/pull/0.log" Oct 02 09:46:13 crc kubenswrapper[4960]: I1002 09:46:13.118516 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/util/0.log" Oct 02 09:46:13 crc kubenswrapper[4960]: I1002 09:46:13.337032 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/extract/0.log" Oct 02 09:46:13 crc kubenswrapper[4960]: I1002 09:46:13.583211 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tbbzq_8a54100f-6654-4909-b7cc-ce0b74ded3f5/marketplace-operator/0.log" Oct 02 09:46:13 crc kubenswrapper[4960]: I1002 09:46:13.807456 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-utilities/0.log" Oct 02 09:46:13 crc kubenswrapper[4960]: I1002 09:46:13.903242 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/registry-server/0.log" Oct 02 09:46:13 crc kubenswrapper[4960]: I1002 09:46:13.996444 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-content/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.047486 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-utilities/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.058668 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-content/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.286538 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-utilities/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.332315 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-content/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.527699 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-utilities/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.603902 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/registry-server/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.769945 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/registry-server/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.781870 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-content/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.808179 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-utilities/0.log" Oct 02 09:46:14 crc kubenswrapper[4960]: I1002 09:46:14.827024 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-content/0.log" Oct 02 09:46:15 crc kubenswrapper[4960]: I1002 09:46:15.012443 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-utilities/0.log" Oct 02 09:46:15 crc kubenswrapper[4960]: I1002 09:46:15.045754 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-content/0.log" Oct 02 09:46:16 crc kubenswrapper[4960]: I1002 09:46:16.193905 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/registry-server/0.log" Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.150181 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.150869 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.150952 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.151796 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"273166d29cffc5df3a7dd810e2a0ba752cb52673b1bea6051b28005b7941d461"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.151865 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://273166d29cffc5df3a7dd810e2a0ba752cb52673b1bea6051b28005b7941d461" gracePeriod=600 Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.966246 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="273166d29cffc5df3a7dd810e2a0ba752cb52673b1bea6051b28005b7941d461" exitCode=0 Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.966807 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"273166d29cffc5df3a7dd810e2a0ba752cb52673b1bea6051b28005b7941d461"} Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.966997 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e"} Oct 02 09:46:29 crc kubenswrapper[4960]: I1002 09:46:29.967027 4960 scope.go:117] "RemoveContainer" containerID="eb36a8e130b036a1143d30754ff10d35b26eb7753aa517485b7e989a86080442" Oct 02 09:47:15 crc kubenswrapper[4960]: E1002 09:47:15.330527 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.281839 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dmxhl"] Oct 02 09:48:07 crc kubenswrapper[4960]: E1002 09:48:07.283171 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="registry-server" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.283189 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="registry-server" Oct 02 09:48:07 crc kubenswrapper[4960]: E1002 09:48:07.283206 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="extract-content" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.283214 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="extract-content" Oct 02 09:48:07 crc kubenswrapper[4960]: E1002 09:48:07.283241 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="extract-utilities" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.283252 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="extract-utilities" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.283517 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="239f1025-ea2f-4d30-81b5-d964fe633cd1" containerName="registry-server" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.285355 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.294004 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmxhl"] Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.360587 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxkpc\" (UniqueName: \"kubernetes.io/projected/6588d8b2-8a6e-4461-ba30-be62e973f8a8-kube-api-access-cxkpc\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.360996 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-catalog-content\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.361238 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-utilities\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.462697 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-catalog-content\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.462777 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-utilities\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.463219 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxkpc\" (UniqueName: \"kubernetes.io/projected/6588d8b2-8a6e-4461-ba30-be62e973f8a8-kube-api-access-cxkpc\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.464361 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-catalog-content\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.464415 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-utilities\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.486764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxkpc\" (UniqueName: \"kubernetes.io/projected/6588d8b2-8a6e-4461-ba30-be62e973f8a8-kube-api-access-cxkpc\") pod \"redhat-operators-dmxhl\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:07 crc kubenswrapper[4960]: I1002 09:48:07.620859 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:08 crc kubenswrapper[4960]: I1002 09:48:08.118389 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmxhl"] Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.014269 4960 generic.go:334] "Generic (PLEG): container finished" podID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerID="1313ce4c7223ffa9682a367870778b991991f4ea96cead78021789a89b832615" exitCode=0 Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.014440 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmxhl" event={"ID":"6588d8b2-8a6e-4461-ba30-be62e973f8a8","Type":"ContainerDied","Data":"1313ce4c7223ffa9682a367870778b991991f4ea96cead78021789a89b832615"} Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.014909 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmxhl" event={"ID":"6588d8b2-8a6e-4461-ba30-be62e973f8a8","Type":"ContainerStarted","Data":"504da7628ed85ace44f2fadce9014bade7187a12eb060e42475715a0cb83ed3d"} Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.017107 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.478800 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t8zhd"] Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.487180 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.492155 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t8zhd"] Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.620384 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwtdf\" (UniqueName: \"kubernetes.io/projected/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-kube-api-access-wwtdf\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.620442 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-catalog-content\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.620615 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-utilities\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.722060 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwtdf\" (UniqueName: \"kubernetes.io/projected/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-kube-api-access-wwtdf\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.722133 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-catalog-content\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.722216 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-utilities\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.722890 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-utilities\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.723619 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-catalog-content\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.752035 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwtdf\" (UniqueName: \"kubernetes.io/projected/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-kube-api-access-wwtdf\") pod \"community-operators-t8zhd\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:09 crc kubenswrapper[4960]: I1002 09:48:09.819080 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:10 crc kubenswrapper[4960]: I1002 09:48:10.377138 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t8zhd"] Oct 02 09:48:11 crc kubenswrapper[4960]: I1002 09:48:11.043301 4960 generic.go:334] "Generic (PLEG): container finished" podID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerID="229416459cc77a67ad6488065bf5ed188bfba61abe2cd62b34142d66178d3867" exitCode=0 Oct 02 09:48:11 crc kubenswrapper[4960]: I1002 09:48:11.043520 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8zhd" event={"ID":"fe7e5f0e-669f-4ca5-a942-046c328d7ad1","Type":"ContainerDied","Data":"229416459cc77a67ad6488065bf5ed188bfba61abe2cd62b34142d66178d3867"} Oct 02 09:48:11 crc kubenswrapper[4960]: I1002 09:48:11.043937 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8zhd" event={"ID":"fe7e5f0e-669f-4ca5-a942-046c328d7ad1","Type":"ContainerStarted","Data":"cc054a5580b6ec3d19bf808cb57c802ee51dbdd1736c608306b3b075f910bf9f"} Oct 02 09:48:11 crc kubenswrapper[4960]: I1002 09:48:11.050159 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmxhl" event={"ID":"6588d8b2-8a6e-4461-ba30-be62e973f8a8","Type":"ContainerStarted","Data":"98110094fb7888e7347bfe5e997bb9d3e2a1c124d1d5fae02a0a0b6fe235563c"} Oct 02 09:48:12 crc kubenswrapper[4960]: I1002 09:48:12.070784 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8zhd" event={"ID":"fe7e5f0e-669f-4ca5-a942-046c328d7ad1","Type":"ContainerStarted","Data":"6321e8b1a8de58cd0b7900c8ee373f10a90a24c8752cc7dc22f989b02d5b66db"} Oct 02 09:48:14 crc kubenswrapper[4960]: I1002 09:48:14.093165 4960 generic.go:334] "Generic (PLEG): container finished" podID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerID="98110094fb7888e7347bfe5e997bb9d3e2a1c124d1d5fae02a0a0b6fe235563c" exitCode=0 Oct 02 09:48:14 crc kubenswrapper[4960]: I1002 09:48:14.093203 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmxhl" event={"ID":"6588d8b2-8a6e-4461-ba30-be62e973f8a8","Type":"ContainerDied","Data":"98110094fb7888e7347bfe5e997bb9d3e2a1c124d1d5fae02a0a0b6fe235563c"} Oct 02 09:48:14 crc kubenswrapper[4960]: I1002 09:48:14.098231 4960 generic.go:334] "Generic (PLEG): container finished" podID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerID="6321e8b1a8de58cd0b7900c8ee373f10a90a24c8752cc7dc22f989b02d5b66db" exitCode=0 Oct 02 09:48:14 crc kubenswrapper[4960]: I1002 09:48:14.098285 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8zhd" event={"ID":"fe7e5f0e-669f-4ca5-a942-046c328d7ad1","Type":"ContainerDied","Data":"6321e8b1a8de58cd0b7900c8ee373f10a90a24c8752cc7dc22f989b02d5b66db"} Oct 02 09:48:15 crc kubenswrapper[4960]: I1002 09:48:15.121080 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8zhd" event={"ID":"fe7e5f0e-669f-4ca5-a942-046c328d7ad1","Type":"ContainerStarted","Data":"afab9cbfab2a2f20b21039582b4ec53f4ccea5f4420c14b7f146951569e71dd0"} Oct 02 09:48:15 crc kubenswrapper[4960]: I1002 09:48:15.130358 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmxhl" event={"ID":"6588d8b2-8a6e-4461-ba30-be62e973f8a8","Type":"ContainerStarted","Data":"b23c6f6234ec5c6b58d025101ab84ebefd1f60c767ee96ad8a4116f908a4e5f8"} Oct 02 09:48:15 crc kubenswrapper[4960]: I1002 09:48:15.157326 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t8zhd" podStartSLOduration=2.570803886 podStartE2EDuration="6.157300181s" podCreationTimestamp="2025-10-02 09:48:09 +0000 UTC" firstStartedPulling="2025-10-02 09:48:11.059019525 +0000 UTC m=+9112.090965812" lastFinishedPulling="2025-10-02 09:48:14.64551582 +0000 UTC m=+9115.677462107" observedRunningTime="2025-10-02 09:48:15.148711252 +0000 UTC m=+9116.180657539" watchObservedRunningTime="2025-10-02 09:48:15.157300181 +0000 UTC m=+9116.189246458" Oct 02 09:48:15 crc kubenswrapper[4960]: I1002 09:48:15.171467 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dmxhl" podStartSLOduration=2.420566036 podStartE2EDuration="8.171444518s" podCreationTimestamp="2025-10-02 09:48:07 +0000 UTC" firstStartedPulling="2025-10-02 09:48:09.016834556 +0000 UTC m=+9110.048780843" lastFinishedPulling="2025-10-02 09:48:14.767713038 +0000 UTC m=+9115.799659325" observedRunningTime="2025-10-02 09:48:15.171407737 +0000 UTC m=+9116.203354034" watchObservedRunningTime="2025-10-02 09:48:15.171444518 +0000 UTC m=+9116.203390805" Oct 02 09:48:17 crc kubenswrapper[4960]: I1002 09:48:17.621617 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:17 crc kubenswrapper[4960]: I1002 09:48:17.622277 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:18 crc kubenswrapper[4960]: I1002 09:48:18.669090 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dmxhl" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="registry-server" probeResult="failure" output=< Oct 02 09:48:18 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 09:48:18 crc kubenswrapper[4960]: > Oct 02 09:48:19 crc kubenswrapper[4960]: I1002 09:48:19.819732 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:19 crc kubenswrapper[4960]: I1002 09:48:19.820419 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:19 crc kubenswrapper[4960]: I1002 09:48:19.882868 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:20 crc kubenswrapper[4960]: I1002 09:48:20.247387 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:20 crc kubenswrapper[4960]: I1002 09:48:20.309008 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t8zhd"] Oct 02 09:48:22 crc kubenswrapper[4960]: I1002 09:48:22.210646 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t8zhd" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="registry-server" containerID="cri-o://afab9cbfab2a2f20b21039582b4ec53f4ccea5f4420c14b7f146951569e71dd0" gracePeriod=2 Oct 02 09:48:25 crc kubenswrapper[4960]: E1002 09:48:25.330798 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:48:27 crc kubenswrapper[4960]: I1002 09:48:27.679793 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:27 crc kubenswrapper[4960]: I1002 09:48:27.750433 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:27 crc kubenswrapper[4960]: I1002 09:48:27.929681 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmxhl"] Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.150459 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.151089 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.388281 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t8zhd_fe7e5f0e-669f-4ca5-a942-046c328d7ad1/registry-server/0.log" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.390733 4960 generic.go:334] "Generic (PLEG): container finished" podID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerID="afab9cbfab2a2f20b21039582b4ec53f4ccea5f4420c14b7f146951569e71dd0" exitCode=137 Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.391635 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8zhd" event={"ID":"fe7e5f0e-669f-4ca5-a942-046c328d7ad1","Type":"ContainerDied","Data":"afab9cbfab2a2f20b21039582b4ec53f4ccea5f4420c14b7f146951569e71dd0"} Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.440847 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t8zhd_fe7e5f0e-669f-4ca5-a942-046c328d7ad1/registry-server/0.log" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.442150 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.559114 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwtdf\" (UniqueName: \"kubernetes.io/projected/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-kube-api-access-wwtdf\") pod \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.559291 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-utilities\") pod \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.559440 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-catalog-content\") pod \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\" (UID: \"fe7e5f0e-669f-4ca5-a942-046c328d7ad1\") " Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.560473 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-utilities" (OuterVolumeSpecName: "utilities") pod "fe7e5f0e-669f-4ca5-a942-046c328d7ad1" (UID: "fe7e5f0e-669f-4ca5-a942-046c328d7ad1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.574297 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-kube-api-access-wwtdf" (OuterVolumeSpecName: "kube-api-access-wwtdf") pod "fe7e5f0e-669f-4ca5-a942-046c328d7ad1" (UID: "fe7e5f0e-669f-4ca5-a942-046c328d7ad1"). InnerVolumeSpecName "kube-api-access-wwtdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.618797 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe7e5f0e-669f-4ca5-a942-046c328d7ad1" (UID: "fe7e5f0e-669f-4ca5-a942-046c328d7ad1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.664937 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.665016 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwtdf\" (UniqueName: \"kubernetes.io/projected/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-kube-api-access-wwtdf\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:29 crc kubenswrapper[4960]: I1002 09:48:29.665032 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7e5f0e-669f-4ca5-a942-046c328d7ad1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.413504 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t8zhd_fe7e5f0e-669f-4ca5-a942-046c328d7ad1/registry-server/0.log" Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.415290 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8zhd" Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.415394 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dmxhl" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="registry-server" containerID="cri-o://b23c6f6234ec5c6b58d025101ab84ebefd1f60c767ee96ad8a4116f908a4e5f8" gracePeriod=2 Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.415532 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8zhd" event={"ID":"fe7e5f0e-669f-4ca5-a942-046c328d7ad1","Type":"ContainerDied","Data":"cc054a5580b6ec3d19bf808cb57c802ee51dbdd1736c608306b3b075f910bf9f"} Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.415580 4960 scope.go:117] "RemoveContainer" containerID="afab9cbfab2a2f20b21039582b4ec53f4ccea5f4420c14b7f146951569e71dd0" Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.440080 4960 scope.go:117] "RemoveContainer" containerID="6321e8b1a8de58cd0b7900c8ee373f10a90a24c8752cc7dc22f989b02d5b66db" Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.460182 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t8zhd"] Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.471127 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t8zhd"] Oct 02 09:48:30 crc kubenswrapper[4960]: I1002 09:48:30.498219 4960 scope.go:117] "RemoveContainer" containerID="229416459cc77a67ad6488065bf5ed188bfba61abe2cd62b34142d66178d3867" Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.434653 4960 generic.go:334] "Generic (PLEG): container finished" podID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerID="b23c6f6234ec5c6b58d025101ab84ebefd1f60c767ee96ad8a4116f908a4e5f8" exitCode=0 Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.434782 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmxhl" event={"ID":"6588d8b2-8a6e-4461-ba30-be62e973f8a8","Type":"ContainerDied","Data":"b23c6f6234ec5c6b58d025101ab84ebefd1f60c767ee96ad8a4116f908a4e5f8"} Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.643166 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.726780 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-utilities\") pod \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.727041 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxkpc\" (UniqueName: \"kubernetes.io/projected/6588d8b2-8a6e-4461-ba30-be62e973f8a8-kube-api-access-cxkpc\") pod \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.727153 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-catalog-content\") pod \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\" (UID: \"6588d8b2-8a6e-4461-ba30-be62e973f8a8\") " Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.729182 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-utilities" (OuterVolumeSpecName: "utilities") pod "6588d8b2-8a6e-4461-ba30-be62e973f8a8" (UID: "6588d8b2-8a6e-4461-ba30-be62e973f8a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.737280 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6588d8b2-8a6e-4461-ba30-be62e973f8a8-kube-api-access-cxkpc" (OuterVolumeSpecName: "kube-api-access-cxkpc") pod "6588d8b2-8a6e-4461-ba30-be62e973f8a8" (UID: "6588d8b2-8a6e-4461-ba30-be62e973f8a8"). InnerVolumeSpecName "kube-api-access-cxkpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.825408 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6588d8b2-8a6e-4461-ba30-be62e973f8a8" (UID: "6588d8b2-8a6e-4461-ba30-be62e973f8a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.830747 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxkpc\" (UniqueName: \"kubernetes.io/projected/6588d8b2-8a6e-4461-ba30-be62e973f8a8-kube-api-access-cxkpc\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.830818 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:31 crc kubenswrapper[4960]: I1002 09:48:31.830837 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6588d8b2-8a6e-4461-ba30-be62e973f8a8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.345462 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" path="/var/lib/kubelet/pods/fe7e5f0e-669f-4ca5-a942-046c328d7ad1/volumes" Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.459722 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmxhl" Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.459606 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmxhl" event={"ID":"6588d8b2-8a6e-4461-ba30-be62e973f8a8","Type":"ContainerDied","Data":"504da7628ed85ace44f2fadce9014bade7187a12eb060e42475715a0cb83ed3d"} Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.461067 4960 scope.go:117] "RemoveContainer" containerID="b23c6f6234ec5c6b58d025101ab84ebefd1f60c767ee96ad8a4116f908a4e5f8" Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.483944 4960 scope.go:117] "RemoveContainer" containerID="98110094fb7888e7347bfe5e997bb9d3e2a1c124d1d5fae02a0a0b6fe235563c" Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.516564 4960 scope.go:117] "RemoveContainer" containerID="1313ce4c7223ffa9682a367870778b991991f4ea96cead78021789a89b832615" Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.519098 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmxhl"] Oct 02 09:48:32 crc kubenswrapper[4960]: I1002 09:48:32.529750 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dmxhl"] Oct 02 09:48:34 crc kubenswrapper[4960]: I1002 09:48:34.343607 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" path="/var/lib/kubelet/pods/6588d8b2-8a6e-4461-ba30-be62e973f8a8/volumes" Oct 02 09:48:59 crc kubenswrapper[4960]: I1002 09:48:59.150457 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:48:59 crc kubenswrapper[4960]: I1002 09:48:59.151491 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:49:15 crc kubenswrapper[4960]: I1002 09:49:15.924496 4960 generic.go:334] "Generic (PLEG): container finished" podID="75cbd459-1ded-44ea-94bf-74218922d643" containerID="d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d" exitCode=0 Oct 02 09:49:15 crc kubenswrapper[4960]: I1002 09:49:15.924614 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" event={"ID":"75cbd459-1ded-44ea-94bf-74218922d643","Type":"ContainerDied","Data":"d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d"} Oct 02 09:49:15 crc kubenswrapper[4960]: I1002 09:49:15.925561 4960 scope.go:117] "RemoveContainer" containerID="d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d" Oct 02 09:49:16 crc kubenswrapper[4960]: I1002 09:49:16.671839 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2k8qc_must-gather-ltjpx_75cbd459-1ded-44ea-94bf-74218922d643/gather/0.log" Oct 02 09:49:25 crc kubenswrapper[4960]: I1002 09:49:25.960318 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2k8qc/must-gather-ltjpx"] Oct 02 09:49:25 crc kubenswrapper[4960]: I1002 09:49:25.961452 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" podUID="75cbd459-1ded-44ea-94bf-74218922d643" containerName="copy" containerID="cri-o://2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de" gracePeriod=2 Oct 02 09:49:25 crc kubenswrapper[4960]: I1002 09:49:25.968722 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2k8qc/must-gather-ltjpx"] Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.471840 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2k8qc_must-gather-ltjpx_75cbd459-1ded-44ea-94bf-74218922d643/copy/0.log" Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.472693 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.606145 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75cbd459-1ded-44ea-94bf-74218922d643-must-gather-output\") pod \"75cbd459-1ded-44ea-94bf-74218922d643\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.606315 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2n86\" (UniqueName: \"kubernetes.io/projected/75cbd459-1ded-44ea-94bf-74218922d643-kube-api-access-b2n86\") pod \"75cbd459-1ded-44ea-94bf-74218922d643\" (UID: \"75cbd459-1ded-44ea-94bf-74218922d643\") " Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.614534 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75cbd459-1ded-44ea-94bf-74218922d643-kube-api-access-b2n86" (OuterVolumeSpecName: "kube-api-access-b2n86") pod "75cbd459-1ded-44ea-94bf-74218922d643" (UID: "75cbd459-1ded-44ea-94bf-74218922d643"). InnerVolumeSpecName "kube-api-access-b2n86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.712557 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2n86\" (UniqueName: \"kubernetes.io/projected/75cbd459-1ded-44ea-94bf-74218922d643-kube-api-access-b2n86\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.821698 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75cbd459-1ded-44ea-94bf-74218922d643-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "75cbd459-1ded-44ea-94bf-74218922d643" (UID: "75cbd459-1ded-44ea-94bf-74218922d643"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:26 crc kubenswrapper[4960]: I1002 09:49:26.917808 4960 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75cbd459-1ded-44ea-94bf-74218922d643-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.079621 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2k8qc_must-gather-ltjpx_75cbd459-1ded-44ea-94bf-74218922d643/copy/0.log" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.085119 4960 generic.go:334] "Generic (PLEG): container finished" podID="75cbd459-1ded-44ea-94bf-74218922d643" containerID="2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de" exitCode=143 Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.085279 4960 scope.go:117] "RemoveContainer" containerID="2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.085491 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2k8qc/must-gather-ltjpx" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.145706 4960 scope.go:117] "RemoveContainer" containerID="d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.328582 4960 scope.go:117] "RemoveContainer" containerID="2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de" Oct 02 09:49:27 crc kubenswrapper[4960]: E1002 09:49:27.329418 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de\": container with ID starting with 2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de not found: ID does not exist" containerID="2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.329458 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de"} err="failed to get container status \"2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de\": rpc error: code = NotFound desc = could not find container \"2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de\": container with ID starting with 2b28d24d30e80ac6e9aa9d8832d5e333cca3af49d2e70836aa277f0c81a552de not found: ID does not exist" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.329506 4960 scope.go:117] "RemoveContainer" containerID="d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d" Oct 02 09:49:27 crc kubenswrapper[4960]: E1002 09:49:27.329823 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d\": container with ID starting with d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d not found: ID does not exist" containerID="d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d" Oct 02 09:49:27 crc kubenswrapper[4960]: I1002 09:49:27.329851 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d"} err="failed to get container status \"d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d\": rpc error: code = NotFound desc = could not find container \"d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d\": container with ID starting with d93855b1cd4457c0a32bb564a714d43b2b28689bf57c6a16ea31c1346e1b602d not found: ID does not exist" Oct 02 09:49:28 crc kubenswrapper[4960]: I1002 09:49:28.352353 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75cbd459-1ded-44ea-94bf-74218922d643" path="/var/lib/kubelet/pods/75cbd459-1ded-44ea-94bf-74218922d643/volumes" Oct 02 09:49:29 crc kubenswrapper[4960]: I1002 09:49:29.149892 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:49:29 crc kubenswrapper[4960]: I1002 09:49:29.150521 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:49:29 crc kubenswrapper[4960]: I1002 09:49:29.150636 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:49:29 crc kubenswrapper[4960]: I1002 09:49:29.151498 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:49:29 crc kubenswrapper[4960]: I1002 09:49:29.151632 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" gracePeriod=600 Oct 02 09:49:29 crc kubenswrapper[4960]: E1002 09:49:29.781180 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:49:30 crc kubenswrapper[4960]: I1002 09:49:30.120926 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" exitCode=0 Oct 02 09:49:30 crc kubenswrapper[4960]: I1002 09:49:30.121026 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e"} Oct 02 09:49:30 crc kubenswrapper[4960]: I1002 09:49:30.121112 4960 scope.go:117] "RemoveContainer" containerID="273166d29cffc5df3a7dd810e2a0ba752cb52673b1bea6051b28005b7941d461" Oct 02 09:49:30 crc kubenswrapper[4960]: I1002 09:49:30.122091 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:49:30 crc kubenswrapper[4960]: E1002 09:49:30.122413 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:49:41 crc kubenswrapper[4960]: I1002 09:49:41.329792 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:49:41 crc kubenswrapper[4960]: E1002 09:49:41.330661 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:49:47 crc kubenswrapper[4960]: E1002 09:49:47.330668 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:49:52 crc kubenswrapper[4960]: I1002 09:49:52.330201 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:49:52 crc kubenswrapper[4960]: E1002 09:49:52.331024 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:50:03 crc kubenswrapper[4960]: I1002 09:50:03.330398 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:50:03 crc kubenswrapper[4960]: E1002 09:50:03.331338 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.149860 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtltj/must-gather-nw58z"] Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.150945 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="extract-content" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.150962 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="extract-content" Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.151034 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cbd459-1ded-44ea-94bf-74218922d643" containerName="gather" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151041 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cbd459-1ded-44ea-94bf-74218922d643" containerName="gather" Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.151067 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="registry-server" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151073 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="registry-server" Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.151090 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cbd459-1ded-44ea-94bf-74218922d643" containerName="copy" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151096 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cbd459-1ded-44ea-94bf-74218922d643" containerName="copy" Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.151114 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="extract-content" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151120 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="extract-content" Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.151136 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="extract-utilities" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151142 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="extract-utilities" Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.151155 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="registry-server" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151161 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="registry-server" Oct 02 09:50:07 crc kubenswrapper[4960]: E1002 09:50:07.151172 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="extract-utilities" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151178 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="extract-utilities" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151371 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cbd459-1ded-44ea-94bf-74218922d643" containerName="gather" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151387 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cbd459-1ded-44ea-94bf-74218922d643" containerName="copy" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151397 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="6588d8b2-8a6e-4461-ba30-be62e973f8a8" containerName="registry-server" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.151412 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7e5f0e-669f-4ca5-a942-046c328d7ad1" containerName="registry-server" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.153168 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.159644 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mtltj"/"default-dockercfg-kdrx7" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.160227 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mtltj"/"openshift-service-ca.crt" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.160223 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mtltj"/"kube-root-ca.crt" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.169681 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mtltj/must-gather-nw58z"] Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.231471 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24906f25-d5a3-4df9-9842-7e3e4c9090cb-must-gather-output\") pod \"must-gather-nw58z\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.231525 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbmps\" (UniqueName: \"kubernetes.io/projected/24906f25-d5a3-4df9-9842-7e3e4c9090cb-kube-api-access-qbmps\") pod \"must-gather-nw58z\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.333962 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24906f25-d5a3-4df9-9842-7e3e4c9090cb-must-gather-output\") pod \"must-gather-nw58z\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.334050 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbmps\" (UniqueName: \"kubernetes.io/projected/24906f25-d5a3-4df9-9842-7e3e4c9090cb-kube-api-access-qbmps\") pod \"must-gather-nw58z\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.334993 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24906f25-d5a3-4df9-9842-7e3e4c9090cb-must-gather-output\") pod \"must-gather-nw58z\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.369734 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbmps\" (UniqueName: \"kubernetes.io/projected/24906f25-d5a3-4df9-9842-7e3e4c9090cb-kube-api-access-qbmps\") pod \"must-gather-nw58z\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.482371 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:50:07 crc kubenswrapper[4960]: I1002 09:50:07.998391 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mtltj/must-gather-nw58z"] Oct 02 09:50:08 crc kubenswrapper[4960]: I1002 09:50:08.538238 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/must-gather-nw58z" event={"ID":"24906f25-d5a3-4df9-9842-7e3e4c9090cb","Type":"ContainerStarted","Data":"9c076fd753acddd5cdde17665c0fea16cd5c880f63d64110c2259210ebdc9e82"} Oct 02 09:50:09 crc kubenswrapper[4960]: I1002 09:50:09.549374 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/must-gather-nw58z" event={"ID":"24906f25-d5a3-4df9-9842-7e3e4c9090cb","Type":"ContainerStarted","Data":"9484e3c02bc49addd4a01878b06c33b25015c17c6a4699861dff4b2963f97ea0"} Oct 02 09:50:09 crc kubenswrapper[4960]: I1002 09:50:09.549773 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/must-gather-nw58z" event={"ID":"24906f25-d5a3-4df9-9842-7e3e4c9090cb","Type":"ContainerStarted","Data":"4a19a232496ec3f7573f6e51d1dcad82d08cd9e9dca40bbfef9a450ac2d6c716"} Oct 02 09:50:09 crc kubenswrapper[4960]: I1002 09:50:09.569389 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mtltj/must-gather-nw58z" podStartSLOduration=2.5693700809999998 podStartE2EDuration="2.569370081s" podCreationTimestamp="2025-10-02 09:50:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:09.564131162 +0000 UTC m=+9230.596077449" watchObservedRunningTime="2025-10-02 09:50:09.569370081 +0000 UTC m=+9230.601316368" Oct 02 09:50:12 crc kubenswrapper[4960]: E1002 09:50:12.319723 4960 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.177:40164->38.102.83.177:37215: write tcp 38.102.83.177:40164->38.102.83.177:37215: write: broken pipe Oct 02 09:50:12 crc kubenswrapper[4960]: I1002 09:50:12.955546 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtltj/crc-debug-rpzr6"] Oct 02 09:50:12 crc kubenswrapper[4960]: I1002 09:50:12.957363 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:12 crc kubenswrapper[4960]: I1002 09:50:12.998605 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd51ce1e-42e0-4507-9159-3f9f24417d18-host\") pod \"crc-debug-rpzr6\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:12 crc kubenswrapper[4960]: I1002 09:50:12.998869 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcldb\" (UniqueName: \"kubernetes.io/projected/cd51ce1e-42e0-4507-9159-3f9f24417d18-kube-api-access-pcldb\") pod \"crc-debug-rpzr6\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:13 crc kubenswrapper[4960]: I1002 09:50:13.100753 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd51ce1e-42e0-4507-9159-3f9f24417d18-host\") pod \"crc-debug-rpzr6\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:13 crc kubenswrapper[4960]: I1002 09:50:13.100874 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcldb\" (UniqueName: \"kubernetes.io/projected/cd51ce1e-42e0-4507-9159-3f9f24417d18-kube-api-access-pcldb\") pod \"crc-debug-rpzr6\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:13 crc kubenswrapper[4960]: I1002 09:50:13.100950 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd51ce1e-42e0-4507-9159-3f9f24417d18-host\") pod \"crc-debug-rpzr6\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:13 crc kubenswrapper[4960]: I1002 09:50:13.123826 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcldb\" (UniqueName: \"kubernetes.io/projected/cd51ce1e-42e0-4507-9159-3f9f24417d18-kube-api-access-pcldb\") pod \"crc-debug-rpzr6\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:13 crc kubenswrapper[4960]: I1002 09:50:13.278920 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:50:13 crc kubenswrapper[4960]: W1002 09:50:13.452628 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd51ce1e_42e0_4507_9159_3f9f24417d18.slice/crio-54fe1c4958f2f15bbdb88adf7339baa6acc22406cd5b8bc2710770a6d0a35319 WatchSource:0}: Error finding container 54fe1c4958f2f15bbdb88adf7339baa6acc22406cd5b8bc2710770a6d0a35319: Status 404 returned error can't find the container with id 54fe1c4958f2f15bbdb88adf7339baa6acc22406cd5b8bc2710770a6d0a35319 Oct 02 09:50:13 crc kubenswrapper[4960]: I1002 09:50:13.586152 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" event={"ID":"cd51ce1e-42e0-4507-9159-3f9f24417d18","Type":"ContainerStarted","Data":"54fe1c4958f2f15bbdb88adf7339baa6acc22406cd5b8bc2710770a6d0a35319"} Oct 02 09:50:14 crc kubenswrapper[4960]: I1002 09:50:14.599263 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" event={"ID":"cd51ce1e-42e0-4507-9159-3f9f24417d18","Type":"ContainerStarted","Data":"bb2235195948b7b897acf144cf74d9c1145c3f6a7b99dc8cccaa33eb281cc124"} Oct 02 09:50:14 crc kubenswrapper[4960]: I1002 09:50:14.631706 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" podStartSLOduration=2.63166511 podStartE2EDuration="2.63166511s" podCreationTimestamp="2025-10-02 09:50:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:14.616423124 +0000 UTC m=+9235.648369411" watchObservedRunningTime="2025-10-02 09:50:14.63166511 +0000 UTC m=+9235.663611397" Oct 02 09:50:15 crc kubenswrapper[4960]: I1002 09:50:15.330193 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:50:15 crc kubenswrapper[4960]: E1002 09:50:15.330407 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:50:27 crc kubenswrapper[4960]: I1002 09:50:27.330508 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:50:27 crc kubenswrapper[4960]: E1002 09:50:27.331401 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:50:38 crc kubenswrapper[4960]: I1002 09:50:38.333271 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:50:38 crc kubenswrapper[4960]: E1002 09:50:38.334171 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:50:52 crc kubenswrapper[4960]: I1002 09:50:52.331264 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:50:52 crc kubenswrapper[4960]: E1002 09:50:52.332443 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:50:59 crc kubenswrapper[4960]: E1002 09:50:59.331448 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:51:03 crc kubenswrapper[4960]: I1002 09:51:03.330635 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:51:03 crc kubenswrapper[4960]: E1002 09:51:03.331835 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:51:05 crc kubenswrapper[4960]: I1002 09:51:05.666956 4960 scope.go:117] "RemoveContainer" containerID="18c80e748dd37b8a7d95d93a5d383d3367a02f6dda48e541291d2948b3e7f3c7" Oct 02 09:51:15 crc kubenswrapper[4960]: I1002 09:51:15.330457 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:51:15 crc kubenswrapper[4960]: E1002 09:51:15.331470 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:51:23 crc kubenswrapper[4960]: I1002 09:51:23.043650 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_baea4578-7ae2-42d3-a07a-ef15c47c886b/ansibletest-ansibletest/0.log" Oct 02 09:51:23 crc kubenswrapper[4960]: I1002 09:51:23.259754 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-69bd96fd88-9gstt_76d6cdce-7551-483b-859c-678805e24895/barbican-api/0.log" Oct 02 09:51:23 crc kubenswrapper[4960]: I1002 09:51:23.475159 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-69bd96fd88-9gstt_76d6cdce-7551-483b-859c-678805e24895/barbican-api-log/0.log" Oct 02 09:51:23 crc kubenswrapper[4960]: I1002 09:51:23.633188 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84cf7698d6-jhmj6_2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7/barbican-keystone-listener/0.log" Oct 02 09:51:24 crc kubenswrapper[4960]: I1002 09:51:24.153868 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bd958db77-xxp6q_cc171097-1fed-426b-9aa6-1c5c905fa5c6/barbican-worker/0.log" Oct 02 09:51:24 crc kubenswrapper[4960]: I1002 09:51:24.352452 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bd958db77-xxp6q_cc171097-1fed-426b-9aa6-1c5c905fa5c6/barbican-worker-log/0.log" Oct 02 09:51:24 crc kubenswrapper[4960]: I1002 09:51:24.437735 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84cf7698d6-jhmj6_2a4c8146-5c6a-4f6e-9bad-94ad46a7fdd7/barbican-keystone-listener-log/0.log" Oct 02 09:51:24 crc kubenswrapper[4960]: I1002 09:51:24.659406 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-c6j9w_9d905943-1348-47e8-81a5-5794bb84f7ff/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:24 crc kubenswrapper[4960]: I1002 09:51:24.874634 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/ceilometer-central-agent/0.log" Oct 02 09:51:24 crc kubenswrapper[4960]: I1002 09:51:24.999026 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/ceilometer-notification-agent/0.log" Oct 02 09:51:25 crc kubenswrapper[4960]: I1002 09:51:25.062909 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/proxy-httpd/0.log" Oct 02 09:51:25 crc kubenswrapper[4960]: I1002 09:51:25.262400 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d1225542-cda6-49d9-bdf3-6bbb7d17e8ac/sg-core/0.log" Oct 02 09:51:25 crc kubenswrapper[4960]: I1002 09:51:25.452305 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-bqtz2_363e150c-f1f7-4cd2-9916-e8724ceb0f9a/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:25 crc kubenswrapper[4960]: I1002 09:51:25.669487 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-w889z_05979476-afd0-404e-87ed-23b6706381f2/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:25 crc kubenswrapper[4960]: I1002 09:51:25.892424 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_913ab46c-3d2d-407c-bd43-6f4b9f9e8f58/cinder-api-log/0.log" Oct 02 09:51:25 crc kubenswrapper[4960]: I1002 09:51:25.980477 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_913ab46c-3d2d-407c-bd43-6f4b9f9e8f58/cinder-api/0.log" Oct 02 09:51:26 crc kubenswrapper[4960]: I1002 09:51:26.381153 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0a9c86d8-b033-44fa-9991-08a9aa629b32/cinder-backup/0.log" Oct 02 09:51:26 crc kubenswrapper[4960]: I1002 09:51:26.382964 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0a9c86d8-b033-44fa-9991-08a9aa629b32/probe/0.log" Oct 02 09:51:26 crc kubenswrapper[4960]: I1002 09:51:26.618843 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d/cinder-scheduler/0.log" Oct 02 09:51:26 crc kubenswrapper[4960]: I1002 09:51:26.737347 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f4c3ac70-8228-472d-90e9-dc3ab6fb6e5d/probe/0.log" Oct 02 09:51:26 crc kubenswrapper[4960]: I1002 09:51:26.893496 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3bebb599-2d7f-4d91-8cf1-eda510fd14f6/cinder-volume/0.log" Oct 02 09:51:26 crc kubenswrapper[4960]: I1002 09:51:26.980271 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3bebb599-2d7f-4d91-8cf1-eda510fd14f6/probe/0.log" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.089702 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6g8jt_947ad109-ebe1-4de0-ad2e-4fa4ae75b238/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.176932 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-gt8pw_7fe7ff43-8e73-408d-bd2d-262e48941ccd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.330301 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:51:27 crc kubenswrapper[4960]: E1002 09:51:27.330606 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.345172 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6785d65565-vfxf8_b169e295-17c4-457d-8832-bb79f85eb5cb/init/0.log" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.539038 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6785d65565-vfxf8_b169e295-17c4-457d-8832-bb79f85eb5cb/init/0.log" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.795443 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6785d65565-vfxf8_b169e295-17c4-457d-8832-bb79f85eb5cb/dnsmasq-dns/0.log" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.847058 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8da35d5f-3f76-4e03-9df7-7d26b23c601d/glance-httpd/0.log" Oct 02 09:51:27 crc kubenswrapper[4960]: I1002 09:51:27.848460 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_8da35d5f-3f76-4e03-9df7-7d26b23c601d/glance-log/0.log" Oct 02 09:51:28 crc kubenswrapper[4960]: I1002 09:51:28.081858 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7bde8f5a-7f1c-4c7d-a52c-ee1269736978/glance-log/0.log" Oct 02 09:51:28 crc kubenswrapper[4960]: I1002 09:51:28.114160 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7bde8f5a-7f1c-4c7d-a52c-ee1269736978/glance-httpd/0.log" Oct 02 09:51:28 crc kubenswrapper[4960]: I1002 09:51:28.423544 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-d66c446d8-rxzk8_35082c42-a42f-4eb7-a12c-91c74a6715f4/horizon/0.log" Oct 02 09:51:28 crc kubenswrapper[4960]: I1002 09:51:28.653371 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_179d0cf9-edfa-4b98-8708-1503790d862e/horizontest-tests-horizontest/0.log" Oct 02 09:51:28 crc kubenswrapper[4960]: I1002 09:51:28.891329 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-zlq7q_5aa32b54-0374-4032-b19f-d1e65557fb95/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:29 crc kubenswrapper[4960]: I1002 09:51:29.123489 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6xk6f_e7f8ad39-e87c-435f-b72b-3c59ac00f2b1/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:29 crc kubenswrapper[4960]: I1002 09:51:29.837860 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323201-k4bsb_20148c75-2594-4fa9-b38e-1fdc97757c0f/keystone-cron/0.log" Oct 02 09:51:30 crc kubenswrapper[4960]: I1002 09:51:30.132203 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-d66c446d8-rxzk8_35082c42-a42f-4eb7-a12c-91c74a6715f4/horizon-log/0.log" Oct 02 09:51:30 crc kubenswrapper[4960]: I1002 09:51:30.291818 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323261-sp279_aa245840-cc66-4671-a12c-00295fe7b5a4/keystone-cron/0.log" Oct 02 09:51:30 crc kubenswrapper[4960]: I1002 09:51:30.506925 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_cc53e302-34cd-4f65-9612-bcb47fc6b238/kube-state-metrics/0.log" Oct 02 09:51:30 crc kubenswrapper[4960]: I1002 09:51:30.753009 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-rtjsc_f2f4ac6d-fe72-4957-bcdb-983b63fa172b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:31 crc kubenswrapper[4960]: I1002 09:51:31.172880 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_5773acec-b7dc-4887-b302-832737da62e9/manila-api-log/0.log" Oct 02 09:51:31 crc kubenswrapper[4960]: I1002 09:51:31.333596 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_5773acec-b7dc-4887-b302-832737da62e9/manila-api/0.log" Oct 02 09:51:31 crc kubenswrapper[4960]: I1002 09:51:31.617598 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_c97772e1-2ae5-425b-8532-be4fbc404d15/probe/0.log" Oct 02 09:51:31 crc kubenswrapper[4960]: I1002 09:51:31.736790 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_c97772e1-2ae5-425b-8532-be4fbc404d15/manila-scheduler/0.log" Oct 02 09:51:32 crc kubenswrapper[4960]: I1002 09:51:32.013580 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-567f974799-pm8bc_ced7f231-e268-4f66-87ae-b32b809bfa6f/keystone-api/0.log" Oct 02 09:51:32 crc kubenswrapper[4960]: I1002 09:51:32.093695 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_29e1c556-f157-4c20-93a7-d0ceecc98439/manila-share/0.log" Oct 02 09:51:32 crc kubenswrapper[4960]: I1002 09:51:32.104446 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_29e1c556-f157-4c20-93a7-d0ceecc98439/probe/0.log" Oct 02 09:51:33 crc kubenswrapper[4960]: I1002 09:51:33.742917 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d8fb8597c-r7rqs_4e13c181-da1b-4da1-9ed7-daa1d05f0f5f/neutron-httpd/0.log" Oct 02 09:51:34 crc kubenswrapper[4960]: I1002 09:51:34.233176 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-vbqcw_9708c6b1-7f2f-4820-9072-12f019a1a731/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:34 crc kubenswrapper[4960]: I1002 09:51:34.365594 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d8fb8597c-r7rqs_4e13c181-da1b-4da1-9ed7-daa1d05f0f5f/neutron-api/0.log" Oct 02 09:51:38 crc kubenswrapper[4960]: I1002 09:51:38.836469 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d889074b-a08e-49cf-8272-2b51fd6a9286/nova-api-log/0.log" Oct 02 09:51:40 crc kubenswrapper[4960]: I1002 09:51:40.358835 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d889074b-a08e-49cf-8272-2b51fd6a9286/nova-api-api/0.log" Oct 02 09:51:40 crc kubenswrapper[4960]: I1002 09:51:40.690258 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e82f1c13-1224-4c25-9050-d55a2ff7c5e6/nova-cell0-conductor-conductor/0.log" Oct 02 09:51:41 crc kubenswrapper[4960]: I1002 09:51:41.368609 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c3bd25c3-7b92-4cc8-a683-3ab41bceccf6/nova-cell1-conductor-conductor/0.log" Oct 02 09:51:41 crc kubenswrapper[4960]: I1002 09:51:41.572400 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_bc95fd9f-83d9-4731-904a-ba493b31656c/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 09:51:42 crc kubenswrapper[4960]: I1002 09:51:42.220141 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-w445z_4a1cec6e-b3b0-48e7-aa5a-8969b06b069f/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:42 crc kubenswrapper[4960]: I1002 09:51:42.332695 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:51:42 crc kubenswrapper[4960]: E1002 09:51:42.333384 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:51:42 crc kubenswrapper[4960]: I1002 09:51:42.593476 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a94097fa-b407-419c-9fad-5078a1f8967f/nova-metadata-log/0.log" Oct 02 09:51:43 crc kubenswrapper[4960]: I1002 09:51:43.924586 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_189dd145-b9b7-4dc5-9fae-a44b359f8748/nova-scheduler-scheduler/0.log" Oct 02 09:51:44 crc kubenswrapper[4960]: I1002 09:51:44.364224 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1ea27c7c-a834-41de-9a07-3f0611c3bc23/mysql-bootstrap/0.log" Oct 02 09:51:44 crc kubenswrapper[4960]: I1002 09:51:44.791193 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1ea27c7c-a834-41de-9a07-3f0611c3bc23/mysql-bootstrap/0.log" Oct 02 09:51:44 crc kubenswrapper[4960]: I1002 09:51:44.970778 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1ea27c7c-a834-41de-9a07-3f0611c3bc23/galera/0.log" Oct 02 09:51:45 crc kubenswrapper[4960]: I1002 09:51:45.327322 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3be247d9-d9ed-49f1-9013-2ad50b6e4ed8/mysql-bootstrap/0.log" Oct 02 09:51:45 crc kubenswrapper[4960]: I1002 09:51:45.480363 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3be247d9-d9ed-49f1-9013-2ad50b6e4ed8/mysql-bootstrap/0.log" Oct 02 09:51:45 crc kubenswrapper[4960]: I1002 09:51:45.608641 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3be247d9-d9ed-49f1-9013-2ad50b6e4ed8/galera/0.log" Oct 02 09:51:45 crc kubenswrapper[4960]: I1002 09:51:45.821140 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_34704ecb-b08d-4f67-b6ed-66141a0f6e74/openstackclient/0.log" Oct 02 09:51:46 crc kubenswrapper[4960]: I1002 09:51:46.098674 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-krnzn_ad314560-608d-471a-a6f6-659b9c0755d6/ovn-controller/0.log" Oct 02 09:51:46 crc kubenswrapper[4960]: I1002 09:51:46.307273 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hmxsf_09ac27ea-acdf-444a-b8b8-0f61332416b1/openstack-network-exporter/0.log" Oct 02 09:51:46 crc kubenswrapper[4960]: I1002 09:51:46.552085 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovsdb-server-init/0.log" Oct 02 09:51:46 crc kubenswrapper[4960]: I1002 09:51:46.765711 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovs-vswitchd/0.log" Oct 02 09:51:46 crc kubenswrapper[4960]: I1002 09:51:46.785836 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovsdb-server-init/0.log" Oct 02 09:51:46 crc kubenswrapper[4960]: I1002 09:51:46.946254 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f9tj4_bff7cbe3-e11d-478e-8ce8-1a03becd8ac8/ovsdb-server/0.log" Oct 02 09:51:47 crc kubenswrapper[4960]: I1002 09:51:47.206364 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gdjnm_ddd1ba07-0eca-48a7-8fd5-64d3c2de7ae3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:47 crc kubenswrapper[4960]: I1002 09:51:47.413160 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_419dd9e4-dbbd-4a58-823b-b987afdf20cb/openstack-network-exporter/0.log" Oct 02 09:51:47 crc kubenswrapper[4960]: I1002 09:51:47.469506 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a94097fa-b407-419c-9fad-5078a1f8967f/nova-metadata-metadata/0.log" Oct 02 09:51:47 crc kubenswrapper[4960]: I1002 09:51:47.487927 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_419dd9e4-dbbd-4a58-823b-b987afdf20cb/ovn-northd/0.log" Oct 02 09:51:47 crc kubenswrapper[4960]: I1002 09:51:47.714074 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_225801bc-ca07-4fa8-8d5c-1d56cfef957b/openstack-network-exporter/0.log" Oct 02 09:51:47 crc kubenswrapper[4960]: I1002 09:51:47.750375 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_225801bc-ca07-4fa8-8d5c-1d56cfef957b/ovsdbserver-nb/0.log" Oct 02 09:51:47 crc kubenswrapper[4960]: I1002 09:51:47.932845 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67a98bd9-8fba-44c5-9229-de8a1053d9d6/openstack-network-exporter/0.log" Oct 02 09:51:48 crc kubenswrapper[4960]: I1002 09:51:48.029069 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_67a98bd9-8fba-44c5-9229-de8a1053d9d6/ovsdbserver-sb/0.log" Oct 02 09:51:48 crc kubenswrapper[4960]: I1002 09:51:48.753963 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43e28122-07de-46d8-b2bd-918405699fd7/setup-container/0.log" Oct 02 09:51:48 crc kubenswrapper[4960]: I1002 09:51:48.869799 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-64f4f7488b-q47mf_9c89938d-910f-4d87-85a5-22bfc7cc24e9/placement-api/0.log" Oct 02 09:51:48 crc kubenswrapper[4960]: I1002 09:51:48.934693 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43e28122-07de-46d8-b2bd-918405699fd7/setup-container/0.log" Oct 02 09:51:49 crc kubenswrapper[4960]: I1002 09:51:49.086727 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-64f4f7488b-q47mf_9c89938d-910f-4d87-85a5-22bfc7cc24e9/placement-log/0.log" Oct 02 09:51:49 crc kubenswrapper[4960]: I1002 09:51:49.148438 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43e28122-07de-46d8-b2bd-918405699fd7/rabbitmq/0.log" Oct 02 09:51:49 crc kubenswrapper[4960]: I1002 09:51:49.320913 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_917895a0-f990-4182-bc76-14bfddefa2a4/setup-container/0.log" Oct 02 09:51:49 crc kubenswrapper[4960]: I1002 09:51:49.628013 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_917895a0-f990-4182-bc76-14bfddefa2a4/setup-container/0.log" Oct 02 09:51:49 crc kubenswrapper[4960]: I1002 09:51:49.668838 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_917895a0-f990-4182-bc76-14bfddefa2a4/rabbitmq/0.log" Oct 02 09:51:49 crc kubenswrapper[4960]: I1002 09:51:49.924932 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-q7wv4_defafea0-a1a3-4e6b-82ce-5b431be9ff5b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:49 crc kubenswrapper[4960]: I1002 09:51:49.981219 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ltc7w_e621ea76-77e5-4ae4-aee0-eb92a4f247a5/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:50 crc kubenswrapper[4960]: I1002 09:51:50.281348 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-phrfr_5066f055-a35b-4ef7-a935-038d15d9838c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:50 crc kubenswrapper[4960]: I1002 09:51:50.475785 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-s7hvl_843c7228-62d1-4839-9459-3de9d174f70a/ssh-known-hosts-edpm-deployment/0.log" Oct 02 09:51:50 crc kubenswrapper[4960]: I1002 09:51:50.753371 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_aed3697d-fab0-4c3e-b5d0-21fb0037a412/tempest-tests-tempest-tests-runner/0.log" Oct 02 09:51:50 crc kubenswrapper[4960]: I1002 09:51:50.867640 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_3cf44a73-42f7-41cc-82e4-f085f154cdbd/tempest-tests-tempest-tests-runner/0.log" Oct 02 09:51:51 crc kubenswrapper[4960]: I1002 09:51:51.001051 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_a15540db-b8e9-444e-9d55-7bcb2b1c4531/test-operator-logs-container/0.log" Oct 02 09:51:51 crc kubenswrapper[4960]: I1002 09:51:51.222955 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_d61db07c-7e9e-4cd3-bf6e-4b5db9b9530f/test-operator-logs-container/0.log" Oct 02 09:51:51 crc kubenswrapper[4960]: I1002 09:51:51.385408 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_06a6e149-1b05-45a2-98b0-f53b766d60eb/test-operator-logs-container/0.log" Oct 02 09:51:51 crc kubenswrapper[4960]: I1002 09:51:51.581777 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_6213b3c4-9214-43d0-9ddf-782b3ec673af/test-operator-logs-container/0.log" Oct 02 09:51:51 crc kubenswrapper[4960]: I1002 09:51:51.811059 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_2b480863-210b-4fb7-8980-e939d5fc9603/tobiko-tests-tobiko/0.log" Oct 02 09:51:51 crc kubenswrapper[4960]: I1002 09:51:51.895024 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_79e4f0e3-919f-44af-b6ae-9bd32f4ff5c8/tobiko-tests-tobiko/0.log" Oct 02 09:51:52 crc kubenswrapper[4960]: I1002 09:51:52.172570 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-rnl7j_5488228f-8fc0-46cb-be89-51d65f7ccff8/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:51:52 crc kubenswrapper[4960]: I1002 09:51:52.889038 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fded14c9-2fd8-46e6-83a9-84c4ca056120/memcached/0.log" Oct 02 09:51:55 crc kubenswrapper[4960]: I1002 09:51:55.330267 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:51:55 crc kubenswrapper[4960]: E1002 09:51:55.330899 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:52:02 crc kubenswrapper[4960]: E1002 09:52:02.330859 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:52:07 crc kubenswrapper[4960]: I1002 09:52:07.330438 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:52:07 crc kubenswrapper[4960]: E1002 09:52:07.331362 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:52:21 crc kubenswrapper[4960]: I1002 09:52:21.331407 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:52:21 crc kubenswrapper[4960]: E1002 09:52:21.332149 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:52:35 crc kubenswrapper[4960]: I1002 09:52:35.330336 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:52:35 crc kubenswrapper[4960]: E1002 09:52:35.331231 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:52:43 crc kubenswrapper[4960]: I1002 09:52:43.161973 4960 generic.go:334] "Generic (PLEG): container finished" podID="cd51ce1e-42e0-4507-9159-3f9f24417d18" containerID="bb2235195948b7b897acf144cf74d9c1145c3f6a7b99dc8cccaa33eb281cc124" exitCode=0 Oct 02 09:52:43 crc kubenswrapper[4960]: I1002 09:52:43.162048 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" event={"ID":"cd51ce1e-42e0-4507-9159-3f9f24417d18","Type":"ContainerDied","Data":"bb2235195948b7b897acf144cf74d9c1145c3f6a7b99dc8cccaa33eb281cc124"} Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.303548 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.354459 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtltj/crc-debug-rpzr6"] Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.358663 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtltj/crc-debug-rpzr6"] Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.410373 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd51ce1e-42e0-4507-9159-3f9f24417d18-host\") pod \"cd51ce1e-42e0-4507-9159-3f9f24417d18\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.410576 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd51ce1e-42e0-4507-9159-3f9f24417d18-host" (OuterVolumeSpecName: "host") pod "cd51ce1e-42e0-4507-9159-3f9f24417d18" (UID: "cd51ce1e-42e0-4507-9159-3f9f24417d18"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.410641 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcldb\" (UniqueName: \"kubernetes.io/projected/cd51ce1e-42e0-4507-9159-3f9f24417d18-kube-api-access-pcldb\") pod \"cd51ce1e-42e0-4507-9159-3f9f24417d18\" (UID: \"cd51ce1e-42e0-4507-9159-3f9f24417d18\") " Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.411280 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd51ce1e-42e0-4507-9159-3f9f24417d18-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.417527 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd51ce1e-42e0-4507-9159-3f9f24417d18-kube-api-access-pcldb" (OuterVolumeSpecName: "kube-api-access-pcldb") pod "cd51ce1e-42e0-4507-9159-3f9f24417d18" (UID: "cd51ce1e-42e0-4507-9159-3f9f24417d18"). InnerVolumeSpecName "kube-api-access-pcldb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:44 crc kubenswrapper[4960]: I1002 09:52:44.514919 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcldb\" (UniqueName: \"kubernetes.io/projected/cd51ce1e-42e0-4507-9159-3f9f24417d18-kube-api-access-pcldb\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.195600 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54fe1c4958f2f15bbdb88adf7339baa6acc22406cd5b8bc2710770a6d0a35319" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.195639 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-rpzr6" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.515055 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtltj/crc-debug-vwqmk"] Oct 02 09:52:45 crc kubenswrapper[4960]: E1002 09:52:45.515517 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd51ce1e-42e0-4507-9159-3f9f24417d18" containerName="container-00" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.515529 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd51ce1e-42e0-4507-9159-3f9f24417d18" containerName="container-00" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.515720 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd51ce1e-42e0-4507-9159-3f9f24417d18" containerName="container-00" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.516562 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.639188 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6e536c-42ff-42c6-83d5-586db257f22f-host\") pod \"crc-debug-vwqmk\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.639336 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t58s\" (UniqueName: \"kubernetes.io/projected/ae6e536c-42ff-42c6-83d5-586db257f22f-kube-api-access-4t58s\") pod \"crc-debug-vwqmk\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.741342 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6e536c-42ff-42c6-83d5-586db257f22f-host\") pod \"crc-debug-vwqmk\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.741503 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6e536c-42ff-42c6-83d5-586db257f22f-host\") pod \"crc-debug-vwqmk\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.741947 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t58s\" (UniqueName: \"kubernetes.io/projected/ae6e536c-42ff-42c6-83d5-586db257f22f-kube-api-access-4t58s\") pod \"crc-debug-vwqmk\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.765999 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t58s\" (UniqueName: \"kubernetes.io/projected/ae6e536c-42ff-42c6-83d5-586db257f22f-kube-api-access-4t58s\") pod \"crc-debug-vwqmk\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:45 crc kubenswrapper[4960]: I1002 09:52:45.835897 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:46 crc kubenswrapper[4960]: I1002 09:52:46.211147 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" event={"ID":"ae6e536c-42ff-42c6-83d5-586db257f22f","Type":"ContainerStarted","Data":"7d9470eb54d540466d8920c794de75d5b393456eb0af21f9dfa2f5682a1898e1"} Oct 02 09:52:46 crc kubenswrapper[4960]: I1002 09:52:46.211583 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" event={"ID":"ae6e536c-42ff-42c6-83d5-586db257f22f","Type":"ContainerStarted","Data":"090ca4fd486c4953e2c9b2cb4ccf2c5191c3e0ca54a314662bfe3fca61199cc1"} Oct 02 09:52:46 crc kubenswrapper[4960]: I1002 09:52:46.229380 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" podStartSLOduration=1.229359513 podStartE2EDuration="1.229359513s" podCreationTimestamp="2025-10-02 09:52:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:52:46.226180968 +0000 UTC m=+9387.258127255" watchObservedRunningTime="2025-10-02 09:52:46.229359513 +0000 UTC m=+9387.261305800" Oct 02 09:52:46 crc kubenswrapper[4960]: I1002 09:52:46.348149 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd51ce1e-42e0-4507-9159-3f9f24417d18" path="/var/lib/kubelet/pods/cd51ce1e-42e0-4507-9159-3f9f24417d18/volumes" Oct 02 09:52:48 crc kubenswrapper[4960]: I1002 09:52:48.237764 4960 generic.go:334] "Generic (PLEG): container finished" podID="ae6e536c-42ff-42c6-83d5-586db257f22f" containerID="7d9470eb54d540466d8920c794de75d5b393456eb0af21f9dfa2f5682a1898e1" exitCode=0 Oct 02 09:52:48 crc kubenswrapper[4960]: I1002 09:52:48.237882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" event={"ID":"ae6e536c-42ff-42c6-83d5-586db257f22f","Type":"ContainerDied","Data":"7d9470eb54d540466d8920c794de75d5b393456eb0af21f9dfa2f5682a1898e1"} Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.329661 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:52:49 crc kubenswrapper[4960]: E1002 09:52:49.329995 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.749527 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.824910 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t58s\" (UniqueName: \"kubernetes.io/projected/ae6e536c-42ff-42c6-83d5-586db257f22f-kube-api-access-4t58s\") pod \"ae6e536c-42ff-42c6-83d5-586db257f22f\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.826013 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6e536c-42ff-42c6-83d5-586db257f22f-host\") pod \"ae6e536c-42ff-42c6-83d5-586db257f22f\" (UID: \"ae6e536c-42ff-42c6-83d5-586db257f22f\") " Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.826182 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae6e536c-42ff-42c6-83d5-586db257f22f-host" (OuterVolumeSpecName: "host") pod "ae6e536c-42ff-42c6-83d5-586db257f22f" (UID: "ae6e536c-42ff-42c6-83d5-586db257f22f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.828279 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae6e536c-42ff-42c6-83d5-586db257f22f-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.837359 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae6e536c-42ff-42c6-83d5-586db257f22f-kube-api-access-4t58s" (OuterVolumeSpecName: "kube-api-access-4t58s") pod "ae6e536c-42ff-42c6-83d5-586db257f22f" (UID: "ae6e536c-42ff-42c6-83d5-586db257f22f"). InnerVolumeSpecName "kube-api-access-4t58s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:49 crc kubenswrapper[4960]: I1002 09:52:49.929781 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t58s\" (UniqueName: \"kubernetes.io/projected/ae6e536c-42ff-42c6-83d5-586db257f22f-kube-api-access-4t58s\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:50 crc kubenswrapper[4960]: I1002 09:52:50.259138 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" event={"ID":"ae6e536c-42ff-42c6-83d5-586db257f22f","Type":"ContainerDied","Data":"090ca4fd486c4953e2c9b2cb4ccf2c5191c3e0ca54a314662bfe3fca61199cc1"} Oct 02 09:52:50 crc kubenswrapper[4960]: I1002 09:52:50.259176 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-vwqmk" Oct 02 09:52:50 crc kubenswrapper[4960]: I1002 09:52:50.259193 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="090ca4fd486c4953e2c9b2cb4ccf2c5191c3e0ca54a314662bfe3fca61199cc1" Oct 02 09:52:58 crc kubenswrapper[4960]: I1002 09:52:58.654671 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtltj/crc-debug-vwqmk"] Oct 02 09:52:58 crc kubenswrapper[4960]: I1002 09:52:58.664551 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtltj/crc-debug-vwqmk"] Oct 02 09:52:59 crc kubenswrapper[4960]: I1002 09:52:59.824735 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mtltj/crc-debug-x9ql4"] Oct 02 09:52:59 crc kubenswrapper[4960]: E1002 09:52:59.825773 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae6e536c-42ff-42c6-83d5-586db257f22f" containerName="container-00" Oct 02 09:52:59 crc kubenswrapper[4960]: I1002 09:52:59.825793 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae6e536c-42ff-42c6-83d5-586db257f22f" containerName="container-00" Oct 02 09:52:59 crc kubenswrapper[4960]: I1002 09:52:59.826054 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae6e536c-42ff-42c6-83d5-586db257f22f" containerName="container-00" Oct 02 09:52:59 crc kubenswrapper[4960]: I1002 09:52:59.826903 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:52:59 crc kubenswrapper[4960]: I1002 09:52:59.922933 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scwt8\" (UniqueName: \"kubernetes.io/projected/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-kube-api-access-scwt8\") pod \"crc-debug-x9ql4\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:52:59 crc kubenswrapper[4960]: I1002 09:52:59.923390 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-host\") pod \"crc-debug-x9ql4\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:00 crc kubenswrapper[4960]: I1002 09:53:00.026028 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scwt8\" (UniqueName: \"kubernetes.io/projected/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-kube-api-access-scwt8\") pod \"crc-debug-x9ql4\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:00 crc kubenswrapper[4960]: I1002 09:53:00.026129 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-host\") pod \"crc-debug-x9ql4\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:00 crc kubenswrapper[4960]: I1002 09:53:00.026281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-host\") pod \"crc-debug-x9ql4\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:00 crc kubenswrapper[4960]: I1002 09:53:00.063415 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scwt8\" (UniqueName: \"kubernetes.io/projected/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-kube-api-access-scwt8\") pod \"crc-debug-x9ql4\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:00 crc kubenswrapper[4960]: I1002 09:53:00.150607 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:00 crc kubenswrapper[4960]: I1002 09:53:00.343376 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae6e536c-42ff-42c6-83d5-586db257f22f" path="/var/lib/kubelet/pods/ae6e536c-42ff-42c6-83d5-586db257f22f/volumes" Oct 02 09:53:00 crc kubenswrapper[4960]: I1002 09:53:00.385452 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-x9ql4" event={"ID":"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0","Type":"ContainerStarted","Data":"6eee8b9cb59f92cbf8cfd430d48224b12ab0b212c782670f8b5655d20d9f77cc"} Oct 02 09:53:01 crc kubenswrapper[4960]: I1002 09:53:01.399799 4960 generic.go:334] "Generic (PLEG): container finished" podID="de301b2d-1ce5-4a5b-a8a1-7296c37cacb0" containerID="95e9bf5723e85c2e4abe9b806036830918e29a2ed0be9e7771cb0f484861b780" exitCode=0 Oct 02 09:53:01 crc kubenswrapper[4960]: I1002 09:53:01.399869 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/crc-debug-x9ql4" event={"ID":"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0","Type":"ContainerDied","Data":"95e9bf5723e85c2e4abe9b806036830918e29a2ed0be9e7771cb0f484861b780"} Oct 02 09:53:01 crc kubenswrapper[4960]: I1002 09:53:01.449907 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtltj/crc-debug-x9ql4"] Oct 02 09:53:01 crc kubenswrapper[4960]: I1002 09:53:01.463458 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtltj/crc-debug-x9ql4"] Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.331213 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:53:02 crc kubenswrapper[4960]: E1002 09:53:02.331610 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.541639 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.684898 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scwt8\" (UniqueName: \"kubernetes.io/projected/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-kube-api-access-scwt8\") pod \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.685311 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-host\") pod \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\" (UID: \"de301b2d-1ce5-4a5b-a8a1-7296c37cacb0\") " Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.685500 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-host" (OuterVolumeSpecName: "host") pod "de301b2d-1ce5-4a5b-a8a1-7296c37cacb0" (UID: "de301b2d-1ce5-4a5b-a8a1-7296c37cacb0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.686048 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.692631 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-kube-api-access-scwt8" (OuterVolumeSpecName: "kube-api-access-scwt8") pod "de301b2d-1ce5-4a5b-a8a1-7296c37cacb0" (UID: "de301b2d-1ce5-4a5b-a8a1-7296c37cacb0"). InnerVolumeSpecName "kube-api-access-scwt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:53:02 crc kubenswrapper[4960]: I1002 09:53:02.788269 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scwt8\" (UniqueName: \"kubernetes.io/projected/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0-kube-api-access-scwt8\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:03 crc kubenswrapper[4960]: I1002 09:53:03.438576 4960 scope.go:117] "RemoveContainer" containerID="95e9bf5723e85c2e4abe9b806036830918e29a2ed0be9e7771cb0f484861b780" Oct 02 09:53:03 crc kubenswrapper[4960]: I1002 09:53:03.438675 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/crc-debug-x9ql4" Oct 02 09:53:03 crc kubenswrapper[4960]: E1002 09:53:03.576636 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde301b2d_1ce5_4a5b_a8a1_7296c37cacb0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde301b2d_1ce5_4a5b_a8a1_7296c37cacb0.slice/crio-6eee8b9cb59f92cbf8cfd430d48224b12ab0b212c782670f8b5655d20d9f77cc\": RecentStats: unable to find data in memory cache]" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.096072 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/util/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.322224 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/pull/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.339873 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/util/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.341094 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/pull/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.351544 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de301b2d-1ce5-4a5b-a8a1-7296c37cacb0" path="/var/lib/kubelet/pods/de301b2d-1ce5-4a5b-a8a1-7296c37cacb0/volumes" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.509431 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/util/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.514872 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/pull/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.553995 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0099f90962ae4d3cc6d942ef2ddab275b3922916e1b59044ce042d59e5lzh2s_b728cb6d-14fc-437a-abe1-836ee912d913/extract/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.761314 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-mfg89_b6b8a229-03cf-4cb0-ab8a-a3133dadfc21/kube-rbac-proxy/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.809399 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-dfpnm_539fe2a2-8989-4fb6-a24e-dee67304e96b/kube-rbac-proxy/0.log" Oct 02 09:53:04 crc kubenswrapper[4960]: I1002 09:53:04.826491 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-mfg89_b6b8a229-03cf-4cb0-ab8a-a3133dadfc21/manager/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.025446 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-dfpnm_539fe2a2-8989-4fb6-a24e-dee67304e96b/manager/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.034402 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-7qjq7_6c71ae76-8a1e-4570-8950-8cf4962169e4/kube-rbac-proxy/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.070400 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-7qjq7_6c71ae76-8a1e-4570-8950-8cf4962169e4/manager/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.221772 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-5mk92_b9dd3b02-9840-4eef-b378-b7888406a57c/kube-rbac-proxy/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.327260 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-5mk92_b9dd3b02-9840-4eef-b378-b7888406a57c/manager/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.449139 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-g9bwh_32ba83c2-5436-4b71-9a95-e552613c6bff/kube-rbac-proxy/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.474573 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-g9bwh_32ba83c2-5436-4b71-9a95-e552613c6bff/manager/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.577577 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-cd9cz_72b0c3f8-2a66-42da-aa83-298bda8f332d/kube-rbac-proxy/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.720754 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-cd9cz_72b0c3f8-2a66-42da-aa83-298bda8f332d/manager/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.823655 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-hbzp7_cf515440-a75b-4f32-8c08-2f108da16b56/kube-rbac-proxy/0.log" Oct 02 09:53:05 crc kubenswrapper[4960]: I1002 09:53:05.968631 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-wm4wc_4bcfb151-7975-4567-9a1f-0cb7dd128696/kube-rbac-proxy/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.038091 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-wm4wc_4bcfb151-7975-4567-9a1f-0cb7dd128696/manager/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.046338 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-hbzp7_cf515440-a75b-4f32-8c08-2f108da16b56/manager/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.253260 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-54crk_8c352774-9d31-4d0a-a1ee-0a96bfbf6239/kube-rbac-proxy/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.349965 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-54crk_8c352774-9d31-4d0a-a1ee-0a96bfbf6239/manager/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.409064 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-hkswg_2af8b0c5-6356-4c12-894e-1bb39e4bf981/kube-rbac-proxy/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.576142 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-kgzd2_5d7077dd-5b52-49af-9dd7-c81f361139ec/kube-rbac-proxy/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.592715 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-hkswg_2af8b0c5-6356-4c12-894e-1bb39e4bf981/manager/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.701544 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-kgzd2_5d7077dd-5b52-49af-9dd7-c81f361139ec/manager/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.855415 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-6lpjs_c0d6bd86-ab92-4f37-be9c-003aab10f910/kube-rbac-proxy/0.log" Oct 02 09:53:06 crc kubenswrapper[4960]: I1002 09:53:06.920703 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-6lpjs_c0d6bd86-ab92-4f37-be9c-003aab10f910/manager/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.060037 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-4vtmz_87d511d4-0cbf-49b2-a2e2-b2e2fc726f60/kube-rbac-proxy/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.176520 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-4vtmz_87d511d4-0cbf-49b2-a2e2-b2e2fc726f60/manager/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.228600 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xhd8b_418a02f4-eb9e-4079-969a-c75f8307b078/kube-rbac-proxy/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.292218 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xhd8b_418a02f4-eb9e-4079-969a-c75f8307b078/manager/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.440884 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-787874f5b776jcd_70d6affd-3067-440a-8b1e-69285828be97/kube-rbac-proxy/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.463199 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-787874f5b776jcd_70d6affd-3067-440a-8b1e-69285828be97/manager/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.502849 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-67698bcd47-nz8r6_f802e6c8-39a5-4acf-b85c-29f235283937/kube-rbac-proxy/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.709481 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-57f4f67bc5-p8hvz_cefbfffd-77e3-4fef-a4b6-b4f94a7cee80/kube-rbac-proxy/0.log" Oct 02 09:53:07 crc kubenswrapper[4960]: I1002 09:53:07.723390 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-57f4f67bc5-p8hvz_cefbfffd-77e3-4fef-a4b6-b4f94a7cee80/operator/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.024073 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-lk5kd_a786a744-833c-40f9-91a7-8b5f312a626c/kube-rbac-proxy/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.099725 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-htqft_12f72b6d-7325-428c-8a30-fd0d75158b31/registry-server/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.129521 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-lk5kd_a786a744-833c-40f9-91a7-8b5f312a626c/manager/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.351107 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-fkgnk_de18430c-640c-46ac-9eb5-de6216332a80/kube-rbac-proxy/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.367339 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-fkgnk_de18430c-640c-46ac-9eb5-de6216332a80/manager/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.597671 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-f8bzk_2fac2998-8d33-448f-877a-388b6570a141/operator/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.632453 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-lzlkk_e52d2a9e-b404-44cf-978a-a2fed39d6337/manager/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.633245 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-lzlkk_e52d2a9e-b404-44cf-978a-a2fed39d6337/kube-rbac-proxy/0.log" Oct 02 09:53:08 crc kubenswrapper[4960]: I1002 09:53:08.862034 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-cmrs5_73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894/kube-rbac-proxy/0.log" Oct 02 09:53:09 crc kubenswrapper[4960]: I1002 09:53:09.065320 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-cmrs5_73c2cbc2-3bd6-44c1-9f5e-ed2bef4a2894/manager/0.log" Oct 02 09:53:09 crc kubenswrapper[4960]: I1002 09:53:09.092692 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-fd7b549cd-82rf9_868e4045-9e39-4add-9023-4ef41caed499/kube-rbac-proxy/0.log" Oct 02 09:53:09 crc kubenswrapper[4960]: I1002 09:53:09.131550 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-67698bcd47-nz8r6_f802e6c8-39a5-4acf-b85c-29f235283937/manager/0.log" Oct 02 09:53:09 crc kubenswrapper[4960]: I1002 09:53:09.188390 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-fd7b549cd-82rf9_868e4045-9e39-4add-9023-4ef41caed499/manager/0.log" Oct 02 09:53:09 crc kubenswrapper[4960]: I1002 09:53:09.302920 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-69k7g_97605d31-7ce2-4f75-848a-1fcf39e54d3e/kube-rbac-proxy/0.log" Oct 02 09:53:09 crc kubenswrapper[4960]: E1002 09:53:09.330586 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:53:09 crc kubenswrapper[4960]: I1002 09:53:09.381421 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-69k7g_97605d31-7ce2-4f75-848a-1fcf39e54d3e/manager/0.log" Oct 02 09:53:17 crc kubenswrapper[4960]: I1002 09:53:17.330762 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:53:17 crc kubenswrapper[4960]: E1002 09:53:17.331846 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:53:29 crc kubenswrapper[4960]: I1002 09:53:29.705214 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4xjvf_3d80dacd-d5ee-4616-878d-0563d69b6b6f/control-plane-machine-set-operator/0.log" Oct 02 09:53:29 crc kubenswrapper[4960]: I1002 09:53:29.872924 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zcpbw_4e0d9ffd-9187-4446-9a5b-87b1b999ff6e/kube-rbac-proxy/0.log" Oct 02 09:53:29 crc kubenswrapper[4960]: I1002 09:53:29.931641 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zcpbw_4e0d9ffd-9187-4446-9a5b-87b1b999ff6e/machine-api-operator/0.log" Oct 02 09:53:30 crc kubenswrapper[4960]: I1002 09:53:30.339169 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:53:30 crc kubenswrapper[4960]: E1002 09:53:30.340084 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:53:42 crc kubenswrapper[4960]: I1002 09:53:42.334096 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:53:42 crc kubenswrapper[4960]: E1002 09:53:42.336970 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:53:42 crc kubenswrapper[4960]: I1002 09:53:42.529776 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vn5rf_aa123932-baa6-435e-b311-e6a62417f39f/cert-manager-controller/0.log" Oct 02 09:53:42 crc kubenswrapper[4960]: I1002 09:53:42.756597 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-74597_c9ca51b7-0180-491c-9d6d-730bea696f09/cert-manager-cainjector/0.log" Oct 02 09:53:42 crc kubenswrapper[4960]: I1002 09:53:42.795085 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-zh4pv_3a34826f-fbd3-4abe-a67f-adc680605fb9/cert-manager-webhook/0.log" Oct 02 09:53:55 crc kubenswrapper[4960]: I1002 09:53:55.124523 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-k8xfq_d3bfd63a-eb72-47a8-a5d1-ca68b6fd40fb/nmstate-console-plugin/0.log" Oct 02 09:53:55 crc kubenswrapper[4960]: I1002 09:53:55.318647 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-r24b6_bf2ea1e7-4cf6-4bc3-be78-37490d9ffc67/nmstate-handler/0.log" Oct 02 09:53:55 crc kubenswrapper[4960]: I1002 09:53:55.330179 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:53:55 crc kubenswrapper[4960]: E1002 09:53:55.330448 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:53:55 crc kubenswrapper[4960]: I1002 09:53:55.387566 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-77kzr_3242a27c-0911-439a-b346-0f921d212cb8/kube-rbac-proxy/0.log" Oct 02 09:53:55 crc kubenswrapper[4960]: I1002 09:53:55.433584 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-77kzr_3242a27c-0911-439a-b346-0f921d212cb8/nmstate-metrics/0.log" Oct 02 09:53:55 crc kubenswrapper[4960]: I1002 09:53:55.581805 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-vpq7s_e1c55993-cf88-4aac-a665-92e26a69a841/nmstate-operator/0.log" Oct 02 09:53:55 crc kubenswrapper[4960]: I1002 09:53:55.689100 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-cjznw_b1cd06a6-a984-40d3-bd87-e81eedce2c7c/nmstate-webhook/0.log" Oct 02 09:54:07 crc kubenswrapper[4960]: I1002 09:54:07.330536 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:54:07 crc kubenswrapper[4960]: E1002 09:54:07.333098 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:54:12 crc kubenswrapper[4960]: I1002 09:54:12.395312 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rd2k2_2d48cc2b-907f-4811-8cb0-779e2799638d/kube-rbac-proxy/0.log" Oct 02 09:54:12 crc kubenswrapper[4960]: I1002 09:54:12.450693 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rd2k2_2d48cc2b-907f-4811-8cb0-779e2799638d/controller/0.log" Oct 02 09:54:12 crc kubenswrapper[4960]: I1002 09:54:12.604447 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:54:12 crc kubenswrapper[4960]: I1002 09:54:12.857565 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:54:12 crc kubenswrapper[4960]: I1002 09:54:12.864170 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:54:12 crc kubenswrapper[4960]: I1002 09:54:12.873950 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:54:12 crc kubenswrapper[4960]: I1002 09:54:12.885003 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.061794 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.062520 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.085208 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.117641 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.332751 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-reloader/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.350406 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-metrics/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.376065 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/cp-frr-files/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.388827 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/controller/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.543772 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/frr-metrics/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.619923 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/kube-rbac-proxy/0.log" Oct 02 09:54:13 crc kubenswrapper[4960]: I1002 09:54:13.643883 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/kube-rbac-proxy-frr/0.log" Oct 02 09:54:14 crc kubenswrapper[4960]: I1002 09:54:14.054367 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/reloader/0.log" Oct 02 09:54:14 crc kubenswrapper[4960]: I1002 09:54:14.147272 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-68xcx_ab470676-e7e9-40a7-96db-3b1cc494dc0a/frr-k8s-webhook-server/0.log" Oct 02 09:54:14 crc kubenswrapper[4960]: I1002 09:54:14.661433 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-cf55b8d8b-rzqpg_b96bb98b-d48b-47d6-8357-0c69cd9052c1/manager/0.log" Oct 02 09:54:14 crc kubenswrapper[4960]: I1002 09:54:14.929176 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-db8949b7d-zx4b9_8d221df4-ed14-4f8d-82db-6326eaa53a3d/webhook-server/0.log" Oct 02 09:54:14 crc kubenswrapper[4960]: I1002 09:54:14.940163 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5j29v_a4cae542-7390-4af8-8c12-97addccc5b4b/kube-rbac-proxy/0.log" Oct 02 09:54:15 crc kubenswrapper[4960]: I1002 09:54:15.813847 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5j29v_a4cae542-7390-4af8-8c12-97addccc5b4b/speaker/0.log" Oct 02 09:54:16 crc kubenswrapper[4960]: I1002 09:54:16.163391 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tr9mf_ab0c4682-a0f3-4f1f-9974-6a5f18e48ed0/frr/0.log" Oct 02 09:54:21 crc kubenswrapper[4960]: I1002 09:54:21.330510 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:54:21 crc kubenswrapper[4960]: E1002 09:54:21.331497 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rfcbg_openshift-machine-config-operator(a813e5e1-abad-433b-a1b3-24cab8119638)\"" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" Oct 02 09:54:29 crc kubenswrapper[4960]: I1002 09:54:29.600484 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/util/0.log" Oct 02 09:54:29 crc kubenswrapper[4960]: I1002 09:54:29.835686 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/pull/0.log" Oct 02 09:54:29 crc kubenswrapper[4960]: I1002 09:54:29.848858 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/util/0.log" Oct 02 09:54:29 crc kubenswrapper[4960]: I1002 09:54:29.902843 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/pull/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.087765 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/extract/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.139922 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/pull/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.154931 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6d6n_d744d009-5615-41b8-a17c-0e290f4d73c6/util/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.283359 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-utilities/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.495198 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-utilities/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.534186 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-content/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.534366 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-content/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.755633 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-content/0.log" Oct 02 09:54:30 crc kubenswrapper[4960]: I1002 09:54:30.756865 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/extract-utilities/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.023159 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-utilities/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.224363 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-content/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.260698 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-content/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.264080 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-utilities/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.577692 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-content/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.602456 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/extract-utilities/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.839519 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/util/0.log" Oct 02 09:54:31 crc kubenswrapper[4960]: I1002 09:54:31.876657 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-drs8j_3a353733-ab76-4e1f-8c4d-9c4dd542e0e9/registry-server/0.log" Oct 02 09:54:32 crc kubenswrapper[4960]: I1002 09:54:32.115436 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/util/0.log" Oct 02 09:54:32 crc kubenswrapper[4960]: I1002 09:54:32.150421 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/pull/0.log" Oct 02 09:54:32 crc kubenswrapper[4960]: I1002 09:54:32.258710 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/pull/0.log" Oct 02 09:54:32 crc kubenswrapper[4960]: I1002 09:54:32.525729 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/util/0.log" Oct 02 09:54:32 crc kubenswrapper[4960]: I1002 09:54:32.574700 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/extract/0.log" Oct 02 09:54:32 crc kubenswrapper[4960]: I1002 09:54:32.582434 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqwk4b_cc5b1383-a1a0-4e12-908e-241ef19995ff/pull/0.log" Oct 02 09:54:32 crc kubenswrapper[4960]: I1002 09:54:32.829391 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tbbzq_8a54100f-6654-4909-b7cc-ce0b74ded3f5/marketplace-operator/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.058441 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j9pkx_e06b264d-7f1e-48b9-a067-1d428f2633f4/registry-server/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.098489 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-utilities/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.264925 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-utilities/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.308203 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-content/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.316564 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-content/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.329744 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.572859 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-content/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.589162 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/extract-utilities/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.668249 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-utilities/0.log" Oct 02 09:54:33 crc kubenswrapper[4960]: I1002 09:54:33.972189 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rq7lt_ae6a6e63-ef7b-4fdf-bf06-994e0651fdb7/registry-server/0.log" Oct 02 09:54:34 crc kubenswrapper[4960]: I1002 09:54:34.150156 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-content/0.log" Oct 02 09:54:34 crc kubenswrapper[4960]: I1002 09:54:34.163870 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-utilities/0.log" Oct 02 09:54:34 crc kubenswrapper[4960]: I1002 09:54:34.178456 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-content/0.log" Oct 02 09:54:34 crc kubenswrapper[4960]: I1002 09:54:34.400783 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"f5e1658f5d260c0f6cd086ee7618db6ba9044835d664d11563a505957829b777"} Oct 02 09:54:34 crc kubenswrapper[4960]: I1002 09:54:34.471321 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-utilities/0.log" Oct 02 09:54:34 crc kubenswrapper[4960]: I1002 09:54:34.480259 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/extract-content/0.log" Oct 02 09:54:35 crc kubenswrapper[4960]: E1002 09:54:35.330709 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:54:35 crc kubenswrapper[4960]: I1002 09:54:35.593373 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nw7n6_95e601cc-5a5e-4630-9693-a7f6c42174d9/registry-server/0.log" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.391779 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tsk67"] Oct 02 09:55:12 crc kubenswrapper[4960]: E1002 09:55:12.392816 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de301b2d-1ce5-4a5b-a8a1-7296c37cacb0" containerName="container-00" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.392828 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="de301b2d-1ce5-4a5b-a8a1-7296c37cacb0" containerName="container-00" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.393036 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="de301b2d-1ce5-4a5b-a8a1-7296c37cacb0" containerName="container-00" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.394459 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.407077 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsk67"] Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.466408 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-utilities\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.466509 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-catalog-content\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.466545 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd2m6\" (UniqueName: \"kubernetes.io/projected/95697a08-ea66-40ab-ad3e-123c27641fb8-kube-api-access-dd2m6\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.568653 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-utilities\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.568751 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-catalog-content\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.568782 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd2m6\" (UniqueName: \"kubernetes.io/projected/95697a08-ea66-40ab-ad3e-123c27641fb8-kube-api-access-dd2m6\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.569291 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-utilities\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.569347 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-catalog-content\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.589632 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd2m6\" (UniqueName: \"kubernetes.io/projected/95697a08-ea66-40ab-ad3e-123c27641fb8-kube-api-access-dd2m6\") pod \"redhat-marketplace-tsk67\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:12 crc kubenswrapper[4960]: I1002 09:55:12.724461 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:13 crc kubenswrapper[4960]: I1002 09:55:13.250520 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsk67"] Oct 02 09:55:13 crc kubenswrapper[4960]: I1002 09:55:13.887845 4960 generic.go:334] "Generic (PLEG): container finished" podID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerID="835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607" exitCode=0 Oct 02 09:55:13 crc kubenswrapper[4960]: I1002 09:55:13.888301 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsk67" event={"ID":"95697a08-ea66-40ab-ad3e-123c27641fb8","Type":"ContainerDied","Data":"835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607"} Oct 02 09:55:13 crc kubenswrapper[4960]: I1002 09:55:13.888338 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsk67" event={"ID":"95697a08-ea66-40ab-ad3e-123c27641fb8","Type":"ContainerStarted","Data":"21fe3f57f4c000c1ef37e5a210f54991ae85fe399fe21658c4383684c910e9e5"} Oct 02 09:55:13 crc kubenswrapper[4960]: I1002 09:55:13.890964 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:55:15 crc kubenswrapper[4960]: I1002 09:55:15.914158 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsk67" event={"ID":"95697a08-ea66-40ab-ad3e-123c27641fb8","Type":"ContainerStarted","Data":"e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0"} Oct 02 09:55:16 crc kubenswrapper[4960]: I1002 09:55:16.942120 4960 generic.go:334] "Generic (PLEG): container finished" podID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerID="e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0" exitCode=0 Oct 02 09:55:16 crc kubenswrapper[4960]: I1002 09:55:16.942224 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsk67" event={"ID":"95697a08-ea66-40ab-ad3e-123c27641fb8","Type":"ContainerDied","Data":"e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0"} Oct 02 09:55:17 crc kubenswrapper[4960]: I1002 09:55:17.957826 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsk67" event={"ID":"95697a08-ea66-40ab-ad3e-123c27641fb8","Type":"ContainerStarted","Data":"90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371"} Oct 02 09:55:22 crc kubenswrapper[4960]: I1002 09:55:22.725422 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:22 crc kubenswrapper[4960]: I1002 09:55:22.727897 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:22 crc kubenswrapper[4960]: I1002 09:55:22.781281 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:22 crc kubenswrapper[4960]: I1002 09:55:22.813081 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tsk67" podStartSLOduration=7.008399614 podStartE2EDuration="10.813009014s" podCreationTimestamp="2025-10-02 09:55:12 +0000 UTC" firstStartedPulling="2025-10-02 09:55:13.890638987 +0000 UTC m=+9534.922585264" lastFinishedPulling="2025-10-02 09:55:17.695248377 +0000 UTC m=+9538.727194664" observedRunningTime="2025-10-02 09:55:17.982896623 +0000 UTC m=+9539.014842920" watchObservedRunningTime="2025-10-02 09:55:22.813009014 +0000 UTC m=+9543.844955301" Oct 02 09:55:23 crc kubenswrapper[4960]: I1002 09:55:23.052880 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:23 crc kubenswrapper[4960]: I1002 09:55:23.108305 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsk67"] Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.024183 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tsk67" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="registry-server" containerID="cri-o://90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371" gracePeriod=2 Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.552252 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.706388 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd2m6\" (UniqueName: \"kubernetes.io/projected/95697a08-ea66-40ab-ad3e-123c27641fb8-kube-api-access-dd2m6\") pod \"95697a08-ea66-40ab-ad3e-123c27641fb8\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.707012 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-utilities\") pod \"95697a08-ea66-40ab-ad3e-123c27641fb8\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.707086 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-catalog-content\") pod \"95697a08-ea66-40ab-ad3e-123c27641fb8\" (UID: \"95697a08-ea66-40ab-ad3e-123c27641fb8\") " Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.709218 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-utilities" (OuterVolumeSpecName: "utilities") pod "95697a08-ea66-40ab-ad3e-123c27641fb8" (UID: "95697a08-ea66-40ab-ad3e-123c27641fb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.716687 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95697a08-ea66-40ab-ad3e-123c27641fb8-kube-api-access-dd2m6" (OuterVolumeSpecName: "kube-api-access-dd2m6") pod "95697a08-ea66-40ab-ad3e-123c27641fb8" (UID: "95697a08-ea66-40ab-ad3e-123c27641fb8"). InnerVolumeSpecName "kube-api-access-dd2m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.727521 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95697a08-ea66-40ab-ad3e-123c27641fb8" (UID: "95697a08-ea66-40ab-ad3e-123c27641fb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.810345 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.810392 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95697a08-ea66-40ab-ad3e-123c27641fb8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:25 crc kubenswrapper[4960]: I1002 09:55:25.810405 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd2m6\" (UniqueName: \"kubernetes.io/projected/95697a08-ea66-40ab-ad3e-123c27641fb8-kube-api-access-dd2m6\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.040152 4960 generic.go:334] "Generic (PLEG): container finished" podID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerID="90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371" exitCode=0 Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.040205 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsk67" event={"ID":"95697a08-ea66-40ab-ad3e-123c27641fb8","Type":"ContainerDied","Data":"90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371"} Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.040235 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsk67" event={"ID":"95697a08-ea66-40ab-ad3e-123c27641fb8","Type":"ContainerDied","Data":"21fe3f57f4c000c1ef37e5a210f54991ae85fe399fe21658c4383684c910e9e5"} Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.040241 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsk67" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.040255 4960 scope.go:117] "RemoveContainer" containerID="90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.069488 4960 scope.go:117] "RemoveContainer" containerID="e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.088244 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsk67"] Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.100028 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsk67"] Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.118103 4960 scope.go:117] "RemoveContainer" containerID="835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.158614 4960 scope.go:117] "RemoveContainer" containerID="90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371" Oct 02 09:55:26 crc kubenswrapper[4960]: E1002 09:55:26.159574 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371\": container with ID starting with 90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371 not found: ID does not exist" containerID="90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.159643 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371"} err="failed to get container status \"90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371\": rpc error: code = NotFound desc = could not find container \"90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371\": container with ID starting with 90cea2e7e1f9b0b1b08f505554d665062c48d29c6710a232d1419bc03caee371 not found: ID does not exist" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.159689 4960 scope.go:117] "RemoveContainer" containerID="e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0" Oct 02 09:55:26 crc kubenswrapper[4960]: E1002 09:55:26.160134 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0\": container with ID starting with e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0 not found: ID does not exist" containerID="e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.160189 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0"} err="failed to get container status \"e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0\": rpc error: code = NotFound desc = could not find container \"e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0\": container with ID starting with e44eee74f49e2a4da2d22c9752c1146697ea984324da5dc3299ba0d5b7ef10f0 not found: ID does not exist" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.160230 4960 scope.go:117] "RemoveContainer" containerID="835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607" Oct 02 09:55:26 crc kubenswrapper[4960]: E1002 09:55:26.160531 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607\": container with ID starting with 835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607 not found: ID does not exist" containerID="835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.160564 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607"} err="failed to get container status \"835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607\": rpc error: code = NotFound desc = could not find container \"835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607\": container with ID starting with 835fdce205e3a934cf80bbb957bd1c37cb4ced586e0610700307ebae61a5d607 not found: ID does not exist" Oct 02 09:55:26 crc kubenswrapper[4960]: I1002 09:55:26.342279 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" path="/var/lib/kubelet/pods/95697a08-ea66-40ab-ad3e-123c27641fb8/volumes" Oct 02 09:56:03 crc kubenswrapper[4960]: E1002 09:56:03.331427 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:56:59 crc kubenswrapper[4960]: I1002 09:56:59.150631 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:56:59 crc kubenswrapper[4960]: I1002 09:56:59.151408 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.249221 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ngf5l"] Oct 02 09:57:00 crc kubenswrapper[4960]: E1002 09:57:00.250800 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="registry-server" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.250816 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="registry-server" Oct 02 09:57:00 crc kubenswrapper[4960]: E1002 09:57:00.250840 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="extract-utilities" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.250847 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="extract-utilities" Oct 02 09:57:00 crc kubenswrapper[4960]: E1002 09:57:00.250860 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="extract-content" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.250866 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="extract-content" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.251072 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="95697a08-ea66-40ab-ad3e-123c27641fb8" containerName="registry-server" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.252434 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.264997 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ngf5l"] Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.336424 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqqxj\" (UniqueName: \"kubernetes.io/projected/486b6e8a-43b5-44c8-8e49-661dbd59fa83-kube-api-access-xqqxj\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.336846 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-utilities\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.337017 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-catalog-content\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.438660 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqqxj\" (UniqueName: \"kubernetes.io/projected/486b6e8a-43b5-44c8-8e49-661dbd59fa83-kube-api-access-xqqxj\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.438729 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-utilities\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.438781 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-catalog-content\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.439328 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-catalog-content\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.439732 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-utilities\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.459810 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqqxj\" (UniqueName: \"kubernetes.io/projected/486b6e8a-43b5-44c8-8e49-661dbd59fa83-kube-api-access-xqqxj\") pod \"certified-operators-ngf5l\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:00 crc kubenswrapper[4960]: I1002 09:57:00.585244 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:01 crc kubenswrapper[4960]: I1002 09:57:01.174553 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ngf5l"] Oct 02 09:57:02 crc kubenswrapper[4960]: I1002 09:57:02.046238 4960 generic.go:334] "Generic (PLEG): container finished" podID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerID="7053cd1ed4211936b96724923b670be7ae7825357cf9a37255117affd610148a" exitCode=0 Oct 02 09:57:02 crc kubenswrapper[4960]: I1002 09:57:02.046558 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngf5l" event={"ID":"486b6e8a-43b5-44c8-8e49-661dbd59fa83","Type":"ContainerDied","Data":"7053cd1ed4211936b96724923b670be7ae7825357cf9a37255117affd610148a"} Oct 02 09:57:02 crc kubenswrapper[4960]: I1002 09:57:02.046584 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngf5l" event={"ID":"486b6e8a-43b5-44c8-8e49-661dbd59fa83","Type":"ContainerStarted","Data":"bf747a8aa90b0787d5e854d2e67f8b393737a325179dd8a223cf60eceb2273f2"} Oct 02 09:57:03 crc kubenswrapper[4960]: I1002 09:57:03.060392 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngf5l" event={"ID":"486b6e8a-43b5-44c8-8e49-661dbd59fa83","Type":"ContainerStarted","Data":"b4fbba76989d7beeb39cbabea178534081b282c4b6016027fd0859c3f59359d6"} Oct 02 09:57:04 crc kubenswrapper[4960]: I1002 09:57:04.072316 4960 generic.go:334] "Generic (PLEG): container finished" podID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerID="b4fbba76989d7beeb39cbabea178534081b282c4b6016027fd0859c3f59359d6" exitCode=0 Oct 02 09:57:04 crc kubenswrapper[4960]: I1002 09:57:04.072366 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngf5l" event={"ID":"486b6e8a-43b5-44c8-8e49-661dbd59fa83","Type":"ContainerDied","Data":"b4fbba76989d7beeb39cbabea178534081b282c4b6016027fd0859c3f59359d6"} Oct 02 09:57:05 crc kubenswrapper[4960]: I1002 09:57:05.086848 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngf5l" event={"ID":"486b6e8a-43b5-44c8-8e49-661dbd59fa83","Type":"ContainerStarted","Data":"ac070adf943778165364bde7e6a426f423b4a071ea85470fc44364c1b921e18c"} Oct 02 09:57:05 crc kubenswrapper[4960]: I1002 09:57:05.115824 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ngf5l" podStartSLOduration=2.372547104 podStartE2EDuration="5.115794936s" podCreationTimestamp="2025-10-02 09:57:00 +0000 UTC" firstStartedPulling="2025-10-02 09:57:02.04824965 +0000 UTC m=+9643.080195937" lastFinishedPulling="2025-10-02 09:57:04.791497482 +0000 UTC m=+9645.823443769" observedRunningTime="2025-10-02 09:57:05.108762539 +0000 UTC m=+9646.140708846" watchObservedRunningTime="2025-10-02 09:57:05.115794936 +0000 UTC m=+9646.147741223" Oct 02 09:57:05 crc kubenswrapper[4960]: I1002 09:57:05.894339 4960 scope.go:117] "RemoveContainer" containerID="bb2235195948b7b897acf144cf74d9c1145c3f6a7b99dc8cccaa33eb281cc124" Oct 02 09:57:10 crc kubenswrapper[4960]: I1002 09:57:10.586426 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:10 crc kubenswrapper[4960]: I1002 09:57:10.587439 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:10 crc kubenswrapper[4960]: I1002 09:57:10.633535 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:11 crc kubenswrapper[4960]: I1002 09:57:11.199374 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:11 crc kubenswrapper[4960]: I1002 09:57:11.258515 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ngf5l"] Oct 02 09:57:13 crc kubenswrapper[4960]: I1002 09:57:13.172326 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ngf5l" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="registry-server" containerID="cri-o://ac070adf943778165364bde7e6a426f423b4a071ea85470fc44364c1b921e18c" gracePeriod=2 Oct 02 09:57:13 crc kubenswrapper[4960]: E1002 09:57:13.330658 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.185286 4960 generic.go:334] "Generic (PLEG): container finished" podID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerID="ac070adf943778165364bde7e6a426f423b4a071ea85470fc44364c1b921e18c" exitCode=0 Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.185382 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngf5l" event={"ID":"486b6e8a-43b5-44c8-8e49-661dbd59fa83","Type":"ContainerDied","Data":"ac070adf943778165364bde7e6a426f423b4a071ea85470fc44364c1b921e18c"} Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.185933 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ngf5l" event={"ID":"486b6e8a-43b5-44c8-8e49-661dbd59fa83","Type":"ContainerDied","Data":"bf747a8aa90b0787d5e854d2e67f8b393737a325179dd8a223cf60eceb2273f2"} Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.185959 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf747a8aa90b0787d5e854d2e67f8b393737a325179dd8a223cf60eceb2273f2" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.309780 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.481776 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-utilities\") pod \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.481868 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqqxj\" (UniqueName: \"kubernetes.io/projected/486b6e8a-43b5-44c8-8e49-661dbd59fa83-kube-api-access-xqqxj\") pod \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.481996 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-catalog-content\") pod \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\" (UID: \"486b6e8a-43b5-44c8-8e49-661dbd59fa83\") " Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.483698 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-utilities" (OuterVolumeSpecName: "utilities") pod "486b6e8a-43b5-44c8-8e49-661dbd59fa83" (UID: "486b6e8a-43b5-44c8-8e49-661dbd59fa83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.484049 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.493238 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/486b6e8a-43b5-44c8-8e49-661dbd59fa83-kube-api-access-xqqxj" (OuterVolumeSpecName: "kube-api-access-xqqxj") pod "486b6e8a-43b5-44c8-8e49-661dbd59fa83" (UID: "486b6e8a-43b5-44c8-8e49-661dbd59fa83"). InnerVolumeSpecName "kube-api-access-xqqxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.585482 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqqxj\" (UniqueName: \"kubernetes.io/projected/486b6e8a-43b5-44c8-8e49-661dbd59fa83-kube-api-access-xqqxj\") on node \"crc\" DevicePath \"\"" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.839191 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "486b6e8a-43b5-44c8-8e49-661dbd59fa83" (UID: "486b6e8a-43b5-44c8-8e49-661dbd59fa83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:57:14 crc kubenswrapper[4960]: I1002 09:57:14.893007 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/486b6e8a-43b5-44c8-8e49-661dbd59fa83-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:57:15 crc kubenswrapper[4960]: I1002 09:57:15.197777 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ngf5l" Oct 02 09:57:15 crc kubenswrapper[4960]: I1002 09:57:15.247051 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ngf5l"] Oct 02 09:57:15 crc kubenswrapper[4960]: I1002 09:57:15.258752 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ngf5l"] Oct 02 09:57:16 crc kubenswrapper[4960]: I1002 09:57:16.343578 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" path="/var/lib/kubelet/pods/486b6e8a-43b5-44c8-8e49-661dbd59fa83/volumes" Oct 02 09:57:26 crc kubenswrapper[4960]: I1002 09:57:26.342758 4960 generic.go:334] "Generic (PLEG): container finished" podID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerID="4a19a232496ec3f7573f6e51d1dcad82d08cd9e9dca40bbfef9a450ac2d6c716" exitCode=0 Oct 02 09:57:26 crc kubenswrapper[4960]: I1002 09:57:26.355639 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mtltj/must-gather-nw58z" event={"ID":"24906f25-d5a3-4df9-9842-7e3e4c9090cb","Type":"ContainerDied","Data":"4a19a232496ec3f7573f6e51d1dcad82d08cd9e9dca40bbfef9a450ac2d6c716"} Oct 02 09:57:26 crc kubenswrapper[4960]: I1002 09:57:26.357050 4960 scope.go:117] "RemoveContainer" containerID="4a19a232496ec3f7573f6e51d1dcad82d08cd9e9dca40bbfef9a450ac2d6c716" Oct 02 09:57:27 crc kubenswrapper[4960]: I1002 09:57:27.074501 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mtltj_must-gather-nw58z_24906f25-d5a3-4df9-9842-7e3e4c9090cb/gather/0.log" Oct 02 09:57:29 crc kubenswrapper[4960]: I1002 09:57:29.149838 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:57:29 crc kubenswrapper[4960]: I1002 09:57:29.150258 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.220115 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mtltj/must-gather-nw58z"] Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.221006 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mtltj/must-gather-nw58z" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerName="copy" containerID="cri-o://9484e3c02bc49addd4a01878b06c33b25015c17c6a4699861dff4b2963f97ea0" gracePeriod=2 Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.229757 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mtltj/must-gather-nw58z"] Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.509644 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mtltj_must-gather-nw58z_24906f25-d5a3-4df9-9842-7e3e4c9090cb/copy/0.log" Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.510890 4960 generic.go:334] "Generic (PLEG): container finished" podID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerID="9484e3c02bc49addd4a01878b06c33b25015c17c6a4699861dff4b2963f97ea0" exitCode=143 Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.696744 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mtltj_must-gather-nw58z_24906f25-d5a3-4df9-9842-7e3e4c9090cb/copy/0.log" Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.697552 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.773523 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24906f25-d5a3-4df9-9842-7e3e4c9090cb-must-gather-output\") pod \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.773796 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbmps\" (UniqueName: \"kubernetes.io/projected/24906f25-d5a3-4df9-9842-7e3e4c9090cb-kube-api-access-qbmps\") pod \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\" (UID: \"24906f25-d5a3-4df9-9842-7e3e4c9090cb\") " Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.781126 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24906f25-d5a3-4df9-9842-7e3e4c9090cb-kube-api-access-qbmps" (OuterVolumeSpecName: "kube-api-access-qbmps") pod "24906f25-d5a3-4df9-9842-7e3e4c9090cb" (UID: "24906f25-d5a3-4df9-9842-7e3e4c9090cb"). InnerVolumeSpecName "kube-api-access-qbmps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.876526 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbmps\" (UniqueName: \"kubernetes.io/projected/24906f25-d5a3-4df9-9842-7e3e4c9090cb-kube-api-access-qbmps\") on node \"crc\" DevicePath \"\"" Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.966649 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24906f25-d5a3-4df9-9842-7e3e4c9090cb-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "24906f25-d5a3-4df9-9842-7e3e4c9090cb" (UID: "24906f25-d5a3-4df9-9842-7e3e4c9090cb"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:57:42 crc kubenswrapper[4960]: I1002 09:57:42.979097 4960 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24906f25-d5a3-4df9-9842-7e3e4c9090cb-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 09:57:43 crc kubenswrapper[4960]: I1002 09:57:43.522260 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mtltj_must-gather-nw58z_24906f25-d5a3-4df9-9842-7e3e4c9090cb/copy/0.log" Oct 02 09:57:43 crc kubenswrapper[4960]: I1002 09:57:43.523103 4960 scope.go:117] "RemoveContainer" containerID="9484e3c02bc49addd4a01878b06c33b25015c17c6a4699861dff4b2963f97ea0" Oct 02 09:57:43 crc kubenswrapper[4960]: I1002 09:57:43.523201 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mtltj/must-gather-nw58z" Oct 02 09:57:43 crc kubenswrapper[4960]: I1002 09:57:43.543246 4960 scope.go:117] "RemoveContainer" containerID="4a19a232496ec3f7573f6e51d1dcad82d08cd9e9dca40bbfef9a450ac2d6c716" Oct 02 09:57:44 crc kubenswrapper[4960]: I1002 09:57:44.342618 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" path="/var/lib/kubelet/pods/24906f25-d5a3-4df9-9842-7e3e4c9090cb/volumes" Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.150636 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.151411 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.151460 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.152304 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5e1658f5d260c0f6cd086ee7618db6ba9044835d664d11563a505957829b777"} pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.152353 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" containerID="cri-o://f5e1658f5d260c0f6cd086ee7618db6ba9044835d664d11563a505957829b777" gracePeriod=600 Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.686662 4960 generic.go:334] "Generic (PLEG): container finished" podID="a813e5e1-abad-433b-a1b3-24cab8119638" containerID="f5e1658f5d260c0f6cd086ee7618db6ba9044835d664d11563a505957829b777" exitCode=0 Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.686753 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerDied","Data":"f5e1658f5d260c0f6cd086ee7618db6ba9044835d664d11563a505957829b777"} Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.687611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" event={"ID":"a813e5e1-abad-433b-a1b3-24cab8119638","Type":"ContainerStarted","Data":"0a7d3e750407a99c08eb8878d1553de9a6461d8f4dde21cc5f2f0e67da40ff03"} Oct 02 09:57:59 crc kubenswrapper[4960]: I1002 09:57:59.687645 4960 scope.go:117] "RemoveContainer" containerID="32900e17cb90e5fed472084ddad849c392e4704b171c319e32f9b5ae7c8b866e" Oct 02 09:58:40 crc kubenswrapper[4960]: E1002 09:58:40.340170 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.649198 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frgkg"] Oct 02 09:58:50 crc kubenswrapper[4960]: E1002 09:58:50.650069 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="extract-utilities" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650081 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="extract-utilities" Oct 02 09:58:50 crc kubenswrapper[4960]: E1002 09:58:50.650100 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="registry-server" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650106 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="registry-server" Oct 02 09:58:50 crc kubenswrapper[4960]: E1002 09:58:50.650121 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerName="gather" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650127 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerName="gather" Oct 02 09:58:50 crc kubenswrapper[4960]: E1002 09:58:50.650139 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerName="copy" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650146 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerName="copy" Oct 02 09:58:50 crc kubenswrapper[4960]: E1002 09:58:50.650179 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="extract-content" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650185 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="extract-content" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650398 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerName="gather" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650410 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="24906f25-d5a3-4df9-9842-7e3e4c9090cb" containerName="copy" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.650424 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="486b6e8a-43b5-44c8-8e49-661dbd59fa83" containerName="registry-server" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.651863 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.670069 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frgkg"] Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.691535 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-catalog-content\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.691607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jttmk\" (UniqueName: \"kubernetes.io/projected/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-kube-api-access-jttmk\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.691996 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-utilities\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.793573 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-utilities\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.793712 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jttmk\" (UniqueName: \"kubernetes.io/projected/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-kube-api-access-jttmk\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.793740 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-catalog-content\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.794317 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-catalog-content\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.794544 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-utilities\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:50 crc kubenswrapper[4960]: I1002 09:58:50.824303 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jttmk\" (UniqueName: \"kubernetes.io/projected/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-kube-api-access-jttmk\") pod \"redhat-operators-frgkg\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:51 crc kubenswrapper[4960]: I1002 09:58:51.039030 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:58:51 crc kubenswrapper[4960]: I1002 09:58:51.596640 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frgkg"] Oct 02 09:58:52 crc kubenswrapper[4960]: I1002 09:58:52.274823 4960 generic.go:334] "Generic (PLEG): container finished" podID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerID="671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c" exitCode=0 Oct 02 09:58:52 crc kubenswrapper[4960]: I1002 09:58:52.274944 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frgkg" event={"ID":"402e1ffd-1640-49e8-ba79-3c3d52bb73a4","Type":"ContainerDied","Data":"671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c"} Oct 02 09:58:52 crc kubenswrapper[4960]: I1002 09:58:52.275221 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frgkg" event={"ID":"402e1ffd-1640-49e8-ba79-3c3d52bb73a4","Type":"ContainerStarted","Data":"f3fb59897e028d44a369021240e0dbaba7ad3bb5facbb97201c8e00417dd1a8a"} Oct 02 09:58:54 crc kubenswrapper[4960]: I1002 09:58:54.306316 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frgkg" event={"ID":"402e1ffd-1640-49e8-ba79-3c3d52bb73a4","Type":"ContainerStarted","Data":"8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25"} Oct 02 09:58:57 crc kubenswrapper[4960]: I1002 09:58:57.339246 4960 generic.go:334] "Generic (PLEG): container finished" podID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerID="8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25" exitCode=0 Oct 02 09:58:57 crc kubenswrapper[4960]: I1002 09:58:57.339328 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frgkg" event={"ID":"402e1ffd-1640-49e8-ba79-3c3d52bb73a4","Type":"ContainerDied","Data":"8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25"} Oct 02 09:58:58 crc kubenswrapper[4960]: I1002 09:58:58.354247 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frgkg" event={"ID":"402e1ffd-1640-49e8-ba79-3c3d52bb73a4","Type":"ContainerStarted","Data":"e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950"} Oct 02 09:58:58 crc kubenswrapper[4960]: I1002 09:58:58.385945 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frgkg" podStartSLOduration=2.8967497 podStartE2EDuration="8.385922496s" podCreationTimestamp="2025-10-02 09:58:50 +0000 UTC" firstStartedPulling="2025-10-02 09:58:52.2769058 +0000 UTC m=+9753.308852087" lastFinishedPulling="2025-10-02 09:58:57.766078596 +0000 UTC m=+9758.798024883" observedRunningTime="2025-10-02 09:58:58.374680876 +0000 UTC m=+9759.406627163" watchObservedRunningTime="2025-10-02 09:58:58.385922496 +0000 UTC m=+9759.417868783" Oct 02 09:59:01 crc kubenswrapper[4960]: I1002 09:59:01.039185 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:59:01 crc kubenswrapper[4960]: I1002 09:59:01.039787 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:59:02 crc kubenswrapper[4960]: I1002 09:59:02.103491 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frgkg" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="registry-server" probeResult="failure" output=< Oct 02 09:59:02 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Oct 02 09:59:02 crc kubenswrapper[4960]: > Oct 02 09:59:06 crc kubenswrapper[4960]: I1002 09:59:06.005591 4960 scope.go:117] "RemoveContainer" containerID="7d9470eb54d540466d8920c794de75d5b393456eb0af21f9dfa2f5682a1898e1" Oct 02 09:59:11 crc kubenswrapper[4960]: I1002 09:59:11.090876 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:59:11 crc kubenswrapper[4960]: I1002 09:59:11.143940 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:59:12 crc kubenswrapper[4960]: I1002 09:59:12.345033 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frgkg"] Oct 02 09:59:12 crc kubenswrapper[4960]: I1002 09:59:12.481250 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frgkg" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="registry-server" containerID="cri-o://e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950" gracePeriod=2 Oct 02 09:59:12 crc kubenswrapper[4960]: I1002 09:59:12.942757 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.078113 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jttmk\" (UniqueName: \"kubernetes.io/projected/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-kube-api-access-jttmk\") pod \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.078227 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-catalog-content\") pod \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.078373 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-utilities\") pod \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\" (UID: \"402e1ffd-1640-49e8-ba79-3c3d52bb73a4\") " Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.080534 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-utilities" (OuterVolumeSpecName: "utilities") pod "402e1ffd-1640-49e8-ba79-3c3d52bb73a4" (UID: "402e1ffd-1640-49e8-ba79-3c3d52bb73a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.087860 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-kube-api-access-jttmk" (OuterVolumeSpecName: "kube-api-access-jttmk") pod "402e1ffd-1640-49e8-ba79-3c3d52bb73a4" (UID: "402e1ffd-1640-49e8-ba79-3c3d52bb73a4"). InnerVolumeSpecName "kube-api-access-jttmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.170834 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "402e1ffd-1640-49e8-ba79-3c3d52bb73a4" (UID: "402e1ffd-1640-49e8-ba79-3c3d52bb73a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.184064 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jttmk\" (UniqueName: \"kubernetes.io/projected/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-kube-api-access-jttmk\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.184102 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.184112 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/402e1ffd-1640-49e8-ba79-3c3d52bb73a4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.499389 4960 generic.go:334] "Generic (PLEG): container finished" podID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerID="e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950" exitCode=0 Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.499478 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frgkg" event={"ID":"402e1ffd-1640-49e8-ba79-3c3d52bb73a4","Type":"ContainerDied","Data":"e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950"} Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.499944 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frgkg" event={"ID":"402e1ffd-1640-49e8-ba79-3c3d52bb73a4","Type":"ContainerDied","Data":"f3fb59897e028d44a369021240e0dbaba7ad3bb5facbb97201c8e00417dd1a8a"} Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.499517 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frgkg" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.499996 4960 scope.go:117] "RemoveContainer" containerID="e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.529058 4960 scope.go:117] "RemoveContainer" containerID="8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.542018 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frgkg"] Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.549337 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frgkg"] Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.562393 4960 scope.go:117] "RemoveContainer" containerID="671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.611199 4960 scope.go:117] "RemoveContainer" containerID="e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950" Oct 02 09:59:13 crc kubenswrapper[4960]: E1002 09:59:13.611957 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950\": container with ID starting with e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950 not found: ID does not exist" containerID="e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.612054 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950"} err="failed to get container status \"e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950\": rpc error: code = NotFound desc = could not find container \"e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950\": container with ID starting with e9165cd28ff2e2af5d05ac38c084ce050825c988d9101ff6eb37e395a440f950 not found: ID does not exist" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.612103 4960 scope.go:117] "RemoveContainer" containerID="8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25" Oct 02 09:59:13 crc kubenswrapper[4960]: E1002 09:59:13.612772 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25\": container with ID starting with 8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25 not found: ID does not exist" containerID="8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.612832 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25"} err="failed to get container status \"8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25\": rpc error: code = NotFound desc = could not find container \"8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25\": container with ID starting with 8a8d657778cffbbc4717db78cafb336dc3373b86dfd43ee3e1c9f78a289f5d25 not found: ID does not exist" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.612878 4960 scope.go:117] "RemoveContainer" containerID="671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c" Oct 02 09:59:13 crc kubenswrapper[4960]: E1002 09:59:13.613571 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c\": container with ID starting with 671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c not found: ID does not exist" containerID="671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c" Oct 02 09:59:13 crc kubenswrapper[4960]: I1002 09:59:13.613704 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c"} err="failed to get container status \"671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c\": rpc error: code = NotFound desc = could not find container \"671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c\": container with ID starting with 671cac3b44728fb03d446fe222031ad6eb0691d5855647bca8bc4263cb3fd73c not found: ID does not exist" Oct 02 09:59:13 crc kubenswrapper[4960]: E1002 09:59:13.720244 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod402e1ffd_1640_49e8_ba79_3c3d52bb73a4.slice\": RecentStats: unable to find data in memory cache]" Oct 02 09:59:14 crc kubenswrapper[4960]: I1002 09:59:14.341281 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" path="/var/lib/kubelet/pods/402e1ffd-1640-49e8-ba79-3c3d52bb73a4/volumes" Oct 02 09:59:52 crc kubenswrapper[4960]: E1002 09:59:52.330607 4960 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 02 09:59:59 crc kubenswrapper[4960]: I1002 09:59:59.149890 4960 patch_prober.go:28] interesting pod/machine-config-daemon-rfcbg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:59:59 crc kubenswrapper[4960]: I1002 09:59:59.151397 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rfcbg" podUID="a813e5e1-abad-433b-a1b3-24cab8119638" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.163944 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7"] Oct 02 10:00:00 crc kubenswrapper[4960]: E1002 10:00:00.165134 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="extract-content" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.165233 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="extract-content" Oct 02 10:00:00 crc kubenswrapper[4960]: E1002 10:00:00.165335 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="registry-server" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.165417 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="registry-server" Oct 02 10:00:00 crc kubenswrapper[4960]: E1002 10:00:00.165476 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="extract-utilities" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.165548 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="extract-utilities" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.165783 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="402e1ffd-1640-49e8-ba79-3c3d52bb73a4" containerName="registry-server" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.166595 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.169286 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.172305 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.181784 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7"] Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.315003 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fce5415-deeb-4c49-aa9c-96d97fb9927c-config-volume\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.315080 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zklfs\" (UniqueName: \"kubernetes.io/projected/3fce5415-deeb-4c49-aa9c-96d97fb9927c-kube-api-access-zklfs\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.315134 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fce5415-deeb-4c49-aa9c-96d97fb9927c-secret-volume\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.417265 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fce5415-deeb-4c49-aa9c-96d97fb9927c-secret-volume\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.417484 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fce5415-deeb-4c49-aa9c-96d97fb9927c-config-volume\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.418568 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fce5415-deeb-4c49-aa9c-96d97fb9927c-config-volume\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.418616 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zklfs\" (UniqueName: \"kubernetes.io/projected/3fce5415-deeb-4c49-aa9c-96d97fb9927c-kube-api-access-zklfs\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.434709 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fce5415-deeb-4c49-aa9c-96d97fb9927c-secret-volume\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.436490 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zklfs\" (UniqueName: \"kubernetes.io/projected/3fce5415-deeb-4c49-aa9c-96d97fb9927c-kube-api-access-zklfs\") pod \"collect-profiles-29323320-fsmw7\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.494400 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:00 crc kubenswrapper[4960]: I1002 10:00:00.984397 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7"] Oct 02 10:00:01 crc kubenswrapper[4960]: I1002 10:00:01.964896 4960 generic.go:334] "Generic (PLEG): container finished" podID="3fce5415-deeb-4c49-aa9c-96d97fb9927c" containerID="bf7b607b0b01073f630d5db0a5ba1817f30ee79d2c94fc37b09df1a2caf0f4c3" exitCode=0 Oct 02 10:00:01 crc kubenswrapper[4960]: I1002 10:00:01.965228 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" event={"ID":"3fce5415-deeb-4c49-aa9c-96d97fb9927c","Type":"ContainerDied","Data":"bf7b607b0b01073f630d5db0a5ba1817f30ee79d2c94fc37b09df1a2caf0f4c3"} Oct 02 10:00:01 crc kubenswrapper[4960]: I1002 10:00:01.965263 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" event={"ID":"3fce5415-deeb-4c49-aa9c-96d97fb9927c","Type":"ContainerStarted","Data":"ffd388d4eec12c9b51d6fc40c6ef7d85eb2c4dfc8875e15917b7ab5415883a0a"} Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.302239 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.493562 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fce5415-deeb-4c49-aa9c-96d97fb9927c-config-volume\") pod \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.493925 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fce5415-deeb-4c49-aa9c-96d97fb9927c-secret-volume\") pod \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.494029 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zklfs\" (UniqueName: \"kubernetes.io/projected/3fce5415-deeb-4c49-aa9c-96d97fb9927c-kube-api-access-zklfs\") pod \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\" (UID: \"3fce5415-deeb-4c49-aa9c-96d97fb9927c\") " Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.495087 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fce5415-deeb-4c49-aa9c-96d97fb9927c-config-volume" (OuterVolumeSpecName: "config-volume") pod "3fce5415-deeb-4c49-aa9c-96d97fb9927c" (UID: "3fce5415-deeb-4c49-aa9c-96d97fb9927c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.499684 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fce5415-deeb-4c49-aa9c-96d97fb9927c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3fce5415-deeb-4c49-aa9c-96d97fb9927c" (UID: "3fce5415-deeb-4c49-aa9c-96d97fb9927c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.500040 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fce5415-deeb-4c49-aa9c-96d97fb9927c-kube-api-access-zklfs" (OuterVolumeSpecName: "kube-api-access-zklfs") pod "3fce5415-deeb-4c49-aa9c-96d97fb9927c" (UID: "3fce5415-deeb-4c49-aa9c-96d97fb9927c"). InnerVolumeSpecName "kube-api-access-zklfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.596702 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fce5415-deeb-4c49-aa9c-96d97fb9927c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.596772 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zklfs\" (UniqueName: \"kubernetes.io/projected/3fce5415-deeb-4c49-aa9c-96d97fb9927c-kube-api-access-zklfs\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.596781 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fce5415-deeb-4c49-aa9c-96d97fb9927c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.984958 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" event={"ID":"3fce5415-deeb-4c49-aa9c-96d97fb9927c","Type":"ContainerDied","Data":"ffd388d4eec12c9b51d6fc40c6ef7d85eb2c4dfc8875e15917b7ab5415883a0a"} Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.985023 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffd388d4eec12c9b51d6fc40c6ef7d85eb2c4dfc8875e15917b7ab5415883a0a" Oct 02 10:00:03 crc kubenswrapper[4960]: I1002 10:00:03.985093 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-fsmw7" Oct 02 10:00:04 crc kubenswrapper[4960]: I1002 10:00:04.382319 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn"] Oct 02 10:00:04 crc kubenswrapper[4960]: I1002 10:00:04.390557 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323275-f7znn"] Oct 02 10:00:06 crc kubenswrapper[4960]: I1002 10:00:06.344024 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24a96346-3d6a-4a46-a0d4-66050b2019cf" path="/var/lib/kubelet/pods/24a96346-3d6a-4a46-a0d4-66050b2019cf/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067446455024465 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067446456017403 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067422761016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067422761015467 5ustar corecore